Very ML
State-of-the-art Machine Learning News Feed
/r/MachineLearning
последний пост 2 часа назад
[R] Controlled LLM Training on Spectral Sphere
[R] Controlled LLM Training on Spectral Sphere [R] Controlled LLM Training on Spectral Sphere

Your request has been blocked due to a network policy.

If you're running a script or application, please register or sign in with your developer credentials here.

Additionally make sure your User-Agent is not empty and is something unique and descriptive and try again.

if you're supplying an alternate User-Agent string, try changing back to default as that can sometimes result in a block.

If you think that we've incorrectly blocked you or you would like to discuss easier ways to get the data you want, please file a ticket here.

2 часа назад @ reddit.com
[D] CUDA Workstation vs Apple Silicon for ML / LLMs
[D] CUDA Workstation vs Apple Silicon for ML / LLMs

Your request has been blocked due to a network policy.

If you're running a script or application, please register or sign in with your developer credentials here.

Additionally make sure your User-Agent is not empty and is something unique and descriptive and try again.

if you're supplying an alternate User-Agent string, try changing back to default as that can sometimes result in a block.

If you think that we've incorrectly blocked you or you would like to discuss easier ways to get the data you want, please file a ticket here.

4 часа назад @ reddit.com
[D] Classification of low resource language using Deep learning
[D] Classification of low resource language using Deep learning

Your request has been blocked due to a network policy.

If you're running a script or application, please register or sign in with your developer credentials here.

Additionally make sure your User-Agent is not empty and is something unique and descriptive and try again.

if you're supplying an alternate User-Agent string, try changing back to default as that can sometimes result in a block.

If you think that we've incorrectly blocked you or you would like to discuss easier ways to get the data you want, please file a ticket here.

10 часов назад @ reddit.com
[R] My team and I have created a system that autonomously creates pufferlib envs. Looking for a compute sponsor
[R] My team and I have created a system that autonomously creates pufferlib envs. Looking for a compute sponsor

Your request has been blocked due to a network policy.

If you're running a script or application, please register or sign in with your developer credentials here.

Additionally make sure your User-Agent is not empty and is something unique and descriptive and try again.

if you're supplying an alternate User-Agent string, try changing back to default as that can sometimes result in a block.

If you think that we've incorrectly blocked you or you would like to discuss easier ways to get the data you want, please file a ticket here.

11 часов назад @ reddit.com
[D] Some of CVPR 2026 Workshops are released
[D] Some of CVPR 2026 Workshops are released

Your request has been blocked due to a network policy.

If you're running a script or application, please register or sign in with your developer credentials here.

Additionally make sure your User-Agent is not empty and is something unique and descriptive and try again.

if you're supplying an alternate User-Agent string, try changing back to default as that can sometimes result in a block.

If you think that we've incorrectly blocked you or you would like to discuss easier ways to get the data you want, please file a ticket here.

12 часов назад @ reddit.com
[D] TMLR timeline question: how long after rebuttal is it normal to wait for a decision?
[D] TMLR timeline question: how long after rebuttal is it normal to wait for a decision?

Your request has been blocked due to a network policy.

If you're running a script or application, please register or sign in with your developer credentials here.

Additionally make sure your User-Agent is not empty and is something unique and descriptive and try again.

if you're supplying an alternate User-Agent string, try changing back to default as that can sometimes result in a block.

If you think that we've incorrectly blocked you or you would like to discuss easier ways to get the data you want, please file a ticket here.

17 часов назад @ reddit.com
[D] Anyone applied to i.AI (incubator for AI) or No 10 Innovation Fellowship?
[D] Anyone applied to i.AI (incubator for AI) or No 10 Innovation Fellowship?

Your request has been blocked due to a network policy.

If you're running a script or application, please register or sign in with your developer credentials here.

Additionally make sure your User-Agent is not empty and is something unique and descriptive and try again.

if you're supplying an alternate User-Agent string, try changing back to default as that can sometimes result in a block.

If you think that we've incorrectly blocked you or you would like to discuss easier ways to get the data you want, please file a ticket here.

17 часов назад @ reddit.com
[R] Why AI Self-Assessment Actually Works: Measuring Knowledge, Not Experience
[R] Why AI Self-Assessment Actually Works: Measuring Knowledge, Not Experience

Your request has been blocked due to a network policy.

If you're running a script or application, please register or sign in with your developer credentials here.

Additionally make sure your User-Agent is not empty and is something unique and descriptive and try again.

if you're supplying an alternate User-Agent string, try changing back to default as that can sometimes result in a block.

If you think that we've incorrectly blocked you or you would like to discuss easier ways to get the data you want, please file a ticket here.

17 часов назад @ reddit.com
[P] Awesome Physical AI – A curated list of academic papers and resources on Physical AI — focusing on VLA models, world models, embodied intelligence, and robotic foundation models.
[P] Awesome Physical AI – A curated list of academic papers and resources on Physical AI — focusing on VLA models, world models, embodied intelligence, and robotic foundation models.

Your request has been blocked due to a network policy.

If you're running a script or application, please register or sign in with your developer credentials here.

Additionally make sure your User-Agent is not empty and is something unique and descriptive and try again.

if you're supplying an alternate User-Agent string, try changing back to default as that can sometimes result in a block.

If you think that we've incorrectly blocked you or you would like to discuss easier ways to get the data you want, please file a ticket here.

18 часов назад @ reddit.com
[P] Semantic caching for LLMs is way harder than it looks - here's what we learned
[P] Semantic caching for LLMs is way harder than it looks - here's what we learned

Your request has been blocked due to a network policy.

If you're running a script or application, please register or sign in with your developer credentials here.

Additionally make sure your User-Agent is not empty and is something unique and descriptive and try again.

if you're supplying an alternate User-Agent string, try changing back to default as that can sometimes result in a block.

If you think that we've incorrectly blocked you or you would like to discuss easier ways to get the data you want, please file a ticket here.

22 часа назад @ reddit.com
[D] I see more people trying to explain mHC than build it
[D] I see more people trying to explain mHC than build it

Your request has been blocked due to a network policy.

If you're running a script or application, please register or sign in with your developer credentials here.

Additionally make sure your User-Agent is not empty and is something unique and descriptive and try again.

if you're supplying an alternate User-Agent string, try changing back to default as that can sometimes result in a block.

If you think that we've incorrectly blocked you or you would like to discuss easier ways to get the data you want, please file a ticket here.

1 day, 2 hours назад @ reddit.com
[R] Vision Transformers with Self-Distilled Registers, NeurIPS 2025
[R] Vision Transformers with Self-Distilled Registers, NeurIPS 2025

Your request has been blocked due to a network policy.

If you're running a script or application, please register or sign in with your developer credentials here.

Additionally make sure your User-Agent is not empty and is something unique and descriptive and try again.

if you're supplying an alternate User-Agent string, try changing back to default as that can sometimes result in a block.

If you think that we've incorrectly blocked you or you would like to discuss easier ways to get the data you want, please file a ticket here.

1 day, 3 hours назад @ reddit.com
[P] Looking for people who are interested in working on a text-minecraft machine learning model
[P] Looking for people who are interested in working on a text-minecraft machine learning model

Your request has been blocked due to a network policy.

If you're running a script or application, please register or sign in with your developer credentials here.

Additionally make sure your User-Agent is not empty and is something unique and descriptive and try again.

if you're supplying an alternate User-Agent string, try changing back to default as that can sometimes result in a block.

If you think that we've incorrectly blocked you or you would like to discuss easier ways to get the data you want, please file a ticket here.

1 day, 4 hours назад @ reddit.com
[R] (DeepSeek) Conditional Memory via Scalable Lookup: A New Axis of Sparsity for Large Language Models
[R] (DeepSeek) Conditional Memory via Scalable Lookup: A New Axis of Sparsity for Large Language Models

Your request has been blocked due to a network policy.

If you're running a script or application, please register or sign in with your developer credentials here.

Additionally make sure your User-Agent is not empty and is something unique and descriptive and try again.

if you're supplying an alternate User-Agent string, try changing back to default as that can sometimes result in a block.

If you think that we've incorrectly blocked you or you would like to discuss easier ways to get the data you want, please file a ticket here.

1 day, 7 hours назад @ reddit.com
[D] Is anyone actually paying for GPU Cluster TCO Consulting? (Because most companies are overpaying by 20%+)
[D] Is anyone actually paying for GPU Cluster TCO Consulting? (Because most companies are overpaying by 20%+)

Your request has been blocked due to a network policy.

If you're running a script or application, please register or sign in with your developer credentials here.

Additionally make sure your User-Agent is not empty and is something unique and descriptive and try again.

if you're supplying an alternate User-Agent string, try changing back to default as that can sometimes result in a block.

If you think that we've incorrectly blocked you or you would like to discuss easier ways to get the data you want, please file a ticket here.

1 day, 9 hours назад @ reddit.com
Towards Data Science
последний пост 1 час назад
Why Human-Centered Data Analytics Matters More Than Ever
Why Human-Centered Data Analytics Matters More Than Ever Why Human-Centered Data Analytics Matters More Than Ever

At its core, human-centered Data Analytics means designing models and metrics with the end-user in mind, not just the business KPI.

Why Human-Centered Data Analytics Is the FutureAs the world becomes more technically sound and business-driven, we as a society have a declining social and behavioral relevance.

How Can You Practice Human-Centered Data Analytics In Your WorkMy inclination toward a human-centered approach is not a newfound love.

Over the years of working as a senior analytics consultant, integrating the Human-Centered approach asked only for some simple, intentional shifts in how I work and here’s how I practice Human-Centered Data Analytics at my workplace.

Human-Centered Data …

1 час назад @ towardsdatascience.com
What Is a Knowledge Graph — and Why It Matters
What Is a Knowledge Graph — and Why It Matters What Is a Knowledge Graph — and Why It Matters

A knowledge graph is a layered knowledge system in which ontologies define meaning, controlled vocabularies catalog entities, and observational data provides evidence—allowing knowledge to accumulate, evolve, and be reasoned over as understanding improves.

An ontology defines classes and the relationships between classes; it is the theory underpinning the knowledge graph.

The knowledge graph isn’t just saying, “Salvarsan is connected to Treponema pallidum.” It is saying “Salvarsan inhibits Treponema pallidum.” It also states that “Treponema pallidum causes syphilis.” These two facts, combined with the logic encoded in the ontology, enable the knowledge graph to infer a new relationship or f…

2 часа назад @ towardsdatascience.com
Glitches in the Attention Matrix
Glitches in the Attention Matrix Glitches in the Attention Matrix

Using Pretrained ViT Models without understanding when high-norm artifacts could impact your project could result in your project failing.

The high-norm artifacts (or attention sinks) they address.

Relationship between ViT High-Norm Artifacts and LLM Attention SinksA phenomenon similar to the ViT high-norm artifacts — attention sinks — were found in LLMs in the StreamingLLM paper (Xiao et al., ICLR 20246).

They find that adding gating does remove the high-norm artifacts, even though the SoftMax attention would still create such artifacts prior to the gating inside the standard scaled-dot product attention (SDPA).

The fixes at least don’t make the performance worse, although the fixes for LL…

4 часа назад @ towardsdatascience.com
Topic Modeling Techniques for 2026: Seeded Modeling, LLM Integration, and Data Summaries
Topic Modeling Techniques for 2026: Seeded Modeling, LLM Integration, and Data Summaries Topic Modeling Techniques for 2026: Seeded Modeling, LLM Integration, and Data Summaries

Human-readable topic names and descriptions are now becoming more and more an expected result of a well-designed topic modelling pipeline.

We will figure out:How to use text prompts to specify what topic models should focus on (i.e., seeded topic models).

It is, in many aspects, very similar to older topic models, as it formulates topic discovery in terms of matrix factorization.

Seeded KeyNMF topic model combines text prompts with the latest topic model to concentrate modelling on a certain problem.

KeyNMF topic model combines text prompts with the latest topic model to concentrate modelling on a certain problem.

5 часов назад @ towardsdatascience.com
An introduction to AWS Bedrock
An introduction to AWS Bedrock An introduction to AWS Bedrock

How do I access Bedrock?

You will be prompted to enter relevant information,$ aws configure AWS Access Key ID [None]: AKIAIOSFODNN7EXAMPLE AWS Secret Access Key [None]: wJalrXUtnFEMI/K7MDENG/bPxRfiCYEXAMPLEKEY Default region name [None]: us-west-2 Default output format [None]:Giving Bedrock access to a modelBack in the day (a few months ago!

aws bedrock list-foundation-modelsThis will return a JSON result set listing various properties of each model, like this.

], "recommended_escalation_target": "IT admin" }SummaryThis article introduced AWS Bedrock, AWS’s managed gateway to foundation large language models, explaining why it exists, how it fits into the broader AWS AI stack, and how to us…

23 часа назад @ towardsdatascience.com
From ‘Dataslows’ to Dataflows: The Gen2 Performance Revolution in Microsoft Fabric
From ‘Dataslows’ to Dataflows: The Gen2 Performance Revolution in Microsoft Fabric From ‘Dataslows’ to Dataflows: The Gen2 Performance Revolution in Microsoft Fabric

Namely, only ADLS Gen2, Fabric Lakehouse, Folder, and Azure Blob Storage connectors can leverage this new feature.

So, nothing really advanced happens here:Image by authorThe same set of operations/transformations has been applied to all three Dataflows Gen2.

Dataflow Gen2 without enhancementsLet’s now do the same thing, but this time using the new Dataflow Gen2.

Image by authorDataflow Gen2 with Modern EvaluatorOk, the moment of truth — let’s now enable the Modern Evaluator for Dataflow Gen2.

However, things are changing rapidly in the Fabric world, and I love how the Fabric Data Integration team makes constant improvements to the product.

1 day, 1 hour назад @ towardsdatascience.com
Under the Uzès Sun: When Historical Data Reveals the Climate Change
Under the Uzès Sun: When Historical Data Reveals the Climate Change Under the Uzès Sun: When Historical Data Reveals the Climate Change

It serves as a blueprint for a classic data challenge: how to architect a high-performance analytical system capable of making sense of decades of historical data applicable to any domain requiring historical vs. current benchmarking.

To keep things clean, I’ve been extracting raw temperature data from the pile of observations we have.

A loaded schema containing more than 500 million rows of French temperature data stretching back to 1780.

) / 2 , FORMAT_STRING=".#" [T_Avg_Period] as avg( [Period], [T_Avg_Daily] ) , FORMAT_STRING=".#" [T_Avg_Diff] as IIF( isEmpty( [T_Avg_Daily] ), null, [T_Avg_Daily] - [T_Avg_Period] ) select [Time].[Months].

Because I need that sweet, sweet historical data…

1 day, 2 hours назад @ towardsdatascience.com
Why Your ML Model Works in Training But Fails in Production
Why Your ML Model Works in Training But Fails in Production Why Your ML Model Works in Training But Fails in Production

My message is simple: most production ML failures are data and time problems, not modeling problems.

Time Travel: An Assumption LeakTime travel is the most common production ML failure I have seen, and also the least discussed in concrete terms.

When we collapse all of that into a single default value, the model does not see a gap.

If this were the only signal available, most teams would conclude that the data pipeline and model inputs remain healthy.

References & Further Reading[1] ROC Curves and AUC (Google Machine Learning Crash Course)https://developers.google.com/machine-learning/crash-course/classification/roc-and-auc[2] Kolmogorov–Smirnov Test (Wikipedia)https://en.wikipedia.org/wiki…

1 day, 4 hours назад @ towardsdatascience.com
How to Maximize Claude Code Effectiveness
How to Maximize Claude Code Effectiveness How to Maximize Claude Code Effectiveness

In this article, I’ll cover why you should consider using Claude Code for coding, highlighting its effectiveness at implementing code with little to no manual review.

Furthermore, I’ll cover some specific techniques I utilize to get the most out of Claude Code, before highlighting some limitations of Claude Code.

Furthermore, I’ll discuss a few specific techniques I utilize to maximize Claude Code effectiveness, including using plan mode, agentic memory, and having an automation mindset.

If I want to code without performing manual reviews of the code, I tend to use Claude Code more and more.

It should also be noted that there exists a Claude Code extension for VS Code, which can be installe…

1 day, 5 hours назад @ towardsdatascience.com
How AI Can Become Your Personal Language Tutor
How AI Can Become Your Personal Language Tutor How AI Can Become Your Personal Language Tutor

Today, generative AI can play that role on your phone or computer, like an AI language tutor you can use any time.

AI For Pronunciation And Oral ComprehensionMy name is Samir, a supply chain professional who struggled with Mandarin during his six-year stay in China.

This is why the first feature of my app is an AI pronunciation coach.

Input of the AI Pronunciation Analysis Agent – (Image by Samir Saci)In this example, I mispronounced the penultimate word.

Questions and illustrations …Teaching material generationAs explained in the previous section, the sentences are also generated using AI.

2 days, 1 hour назад @ towardsdatascience.com
Why 90% Accuracy in Text-to-SQL is 100% Useless
Why 90% Accuracy in Text-to-SQL is 100% Useless Why 90% Accuracy in Text-to-SQL is 100% Useless

Spider 2.0 simulates this by providing external files (Markdown, YAML) the model must read to ground its reasoning.

The Agentic Workflow: Crucially, Spider 2.0 models the workflow of a modern data engineer.

Achieving 90% accuracy might be academically interesting, but in the enterprise, it is industrially useless.

Only by testing against the messy reality of enterprise data can we develop Text-to-SQL applications reliable enough to bet the business on.

Further ReadingSpider 2.0: Evaluating Language Models on Real-World Enterprise Text-to-SQL Workflows Authors: Fangyu Lei, Jixuan Chen, Yuxiao Ye, et al.

2 days, 2 hours назад @ towardsdatascience.com
When Does Adding Fancy RAG Features Work?
When Does Adding Fancy RAG Features Work? When Does Adding Fancy RAG Features Work?

It penalized the complex pipeline for more peripheral details (edge cases, extra citations) that weren’t directly asked for.

I.e., questions with specific names in them (like “how does CLAUSE compare…”) matched documents fine, and the query optimizer just made things worse.

The judge favored naive’s conciseness and brevity, while it favored the complex pipeline for completeness and comprehensiveness.

In short: the naive answer hallucinates by necessity due to missing context, while the complex answer is verbose but materially supported.

Before we move on to the cost/latency analysis, we can try to isolate the query optimizer as well.

2 days, 4 hours назад @ towardsdatascience.com
Optimizing Data Transfer in Batched AI/ML Inference Workloads
Optimizing Data Transfer in Batched AI/ML Inference Workloads Optimizing Data Transfer in Batched AI/ML Inference Workloads

Moreover, while optimizations to CPU-to-GPU data-loading are well documented and easy to implement, optimizing data copy in the opposite direction requires a bit more manual labor.

We modify our inference loop to put the output buffers in the output queue.

In our previous post, the same finding informed our asynchronous data copy optimization.

Presumably, the large GPU-to-CPU data copy involved quite a bit of memory-management overhead that we have removed by implementing a dedicated buffer pool.

The monitor thread waits for the event to trigger and then pushes the output tensor to the output queue for processing.

2 days, 5 hours назад @ towardsdatascience.com
Automatic Prompt Optimization for Multimodal Vision Agents: A Self-Driving Car Example
Automatic Prompt Optimization for Multimodal Vision Agents: A Self-Driving Car Example Automatic Prompt Optimization for Multimodal Vision Agents: A Self-Driving Car Example

Automatic prompt optimization is a recent advancement in the field that systematically tunes prompts to produce more accurate, consistent outputs.

To best demonstrate how these concepts work in a more complex, dynamic, and data-intensive setting, we will walk through an example using a self-driving car agent.

Agent optimization is part of automatic prompt engineering, but it involves working with various parts of the agent, such as multi-prompts, tool calling, RAG, agent architecture, and various modalities.

Getting StartedIn this walkthrough, we want to use a small dataset of self-driving car dashcam images and tune the prompts using automatic prompt optimization with a multi-modal agent t…

3 days, 2 hours назад @ towardsdatascience.com
How to Leverage Slash Commands to Code Effectively
How to Leverage Slash Commands to Code Effectively How to Leverage Slash Commands to Code Effectively

I’ll discuss what slash commands are, and how you can leverage them to saves a lot of time on a daily basis.

An additional benefit of using slash commands is that your commands will be consistent.

How to make slash commandsYou create slash commands in different ways, depending on which tool you are using.

Generalize the knowledge from a threadAnother command command I use it to generalize knowledge from a thread.

You can see the code review slash command below:# Code Review Checklist ## Overview Comprehensive checklist for conducting thorough code reviews to ensure quality, security, and maintainability.

3 days, 4 hours назад @ towardsdatascience.com
Distill.pub Distill.pub
последний пост None
TheSequence TheSequence
последний пост 5 часов назад
The Sequence AI of the Week #789: Recursive Language Models: Inside the MIT Research Everyone is Talking About
The Sequence AI of the Week #789: Recursive Language Models: Inside the MIT Research Everyone is Talking About The Sequence AI of the Week #789: Recursive Language Models: Inside the MIT Research Everyone is Talking About

Like it skimmed the whole thing with sleepy eyes and then improvised.

The usual remedy has been to fight length with length.

Past a point, the model isn’t reasoning over your input; it’s drowning in it.

The paper names the failure mode plainly: context rot.

Not merely because we hit a hard context limit, but because even within that limit the model’s effective “working set” becomes unmanageably large.

5 часов назад @ thesequence.substack.com
The Sequence Knowledge #788: Inside the Generator: Meet The Top Synthetic Data Generation Frameworks for Modern AI
The Sequence Knowledge #788: Inside the Generator: Meet The Top Synthetic Data Generation Frameworks for Modern AI The Sequence Knowledge #788: Inside the Generator: Meet The Top Synthetic Data Generation Frameworks for Modern AI

Today we will Discuss:An overivew of the top synthetic data generation frameworks in the market.

NVIDIA’s top framework for synthetic data generation.

💡 AI Concept of the Day: An Overview of Synthetic Data Generation FrameworksSynthetic data has quietly become the “second scaling law” for foundation models: once you’ve saturated human-authored corpora, the only way to keep climbing is to manufacture new data with models themselves.

The interesting part is that this is no longer done with ad-hoc scripts; we’re seeing full-fledged frameworks that treat synthetic generation as an infrastructure problem.

NVIDIA: Nemotron-4 + NeMo as a synthetic data foundry

1 day, 5 hours назад @ thesequence.substack.com
The Sequence Radar #787: Rubin, Raises, and Returns: 2026 Starts Fast
The Sequence Radar #787: Rubin, Raises, and Returns: 2026 Starts Fast The Sequence Radar #787: Rubin, Raises, and Returns: 2026 Starts Fast

The AI of the week discusses MIT’s crazy paper about recurvise language models that everyone is talking about.

Subscribe and don’t miss out:📝 Editorial: Rubin, Raises, and Returns: 2026 Starts FastThe first full week of 2026 has been incredibly active, marked by a density of significant developments that sets a high bar for the rest of the year.

The Silicon Wars: From Chips to “AI Factories”At CES 2026, the pretense that the GPU is a standalone component finally ended.

This approach allows the model to achieve state-of-the-art performance on reasoning-heavy benchmarks while significantly reducing computational costs and latency on simpler perception tasks.

🤖 AI Tech ReleasesLFM2.5Liquid AI …

3 days, 5 hours назад @ thesequence.substack.com
The Sequence Opinion #786: The Great Absorption: When System Code Becomes Model Weights
The Sequence Opinion #786: The Great Absorption: When System Code Becomes Model Weights The Sequence Opinion #786: The Great Absorption: When System Code Becomes Model Weights

There is a distinct, unsettling feeling when you write code for AI agents today.

It is a specific flavor of Rich Sutton’s “The Bitter Lesson,” playing out in real-time.

In Sutton’s original formulation, hand-coded features were consistently crushed by general methods that leveraged computation.1 In our current era, the lesson is slightly different but equally bitter: Hand-coded system scaffolding is consistently crushed by model internalization.

The history of the last three years is the history of capabilities migrating from the “outside” (the system, the prompt, the agent loop) to the “inside” (the weights, the activations, the forward pass).

You need to know which parts of your stack are…

6 days, 5 hours назад @ thesequence.substack.com
The Sequence AI of the Week #785: Gradient Highway Maintenance: Inside DeepSeek’s Latest Breakthrough
The Sequence AI of the Week #785: Gradient Highway Maintenance: Inside DeepSeek’s Latest Breakthrough The Sequence AI of the Week #785: Gradient Highway Maintenance: Inside DeepSeek’s Latest Breakthrough

Today, I would like to discuss DeepSeek’s new paper that everyone is talking about.

We love to talk about the flashy stuff—the Attention mechanism, the Mixture-of-Experts routing, the sheer scale of the datasets.

But if you peel back the layers of a Transformer, or really almost any deep network from the last decade, you find the true unsung hero of the Deep Learning revolution: the Residual Connection.

), the residual connection is beautifully simple: y = f(x) + x.

You take the input x, pass it through some non-linear transformation f(x), and then just add the original x back in.

1 week назад @ thesequence.substack.com
The Sequence Knowledge #784: The Convergence of Synthetic Data and World Models Models Are Unlocking Embodied AI
The Sequence Knowledge #784: The Convergence of Synthetic Data and World Models Models Are Unlocking Embodied AI The Sequence Knowledge #784: The Convergence of Synthetic Data and World Models Models Are Unlocking Embodied AI

Today we will Discuss:Synthetic data generation for 3D world model environments.

The amazing Genie world model from Google DeepMind.

Gathering high-fidelity, perfectly labeled 3D data from the physical world is slow, expensive, and rarely captures the “long-tail” edge cases necessary for robust systems.

The industry is increasingly turning to a powerful pairing to solve this: Synthetic Data Generation (SDG) and World Models.

Together, they are moving us from training AI on static datasets to training them inside dynamic, interactive simulations.

1 week, 1 day назад @ thesequence.substack.com
The Sequence Radar #783: Softbank, DeepSeek, MiniMax and The Sequence 2026
The Sequence Radar #783: Softbank, DeepSeek, MiniMax and The Sequence 2026 The Sequence Radar #783: Softbank, DeepSeek, MiniMax and The Sequence 2026

Next Week in The Sequence:We continue our series about synthetic data exploring the potential for world models.

Subscribe and don’t miss out:📝 Editorial: Softbank, DeepSeek, MiniMax and The Sequence 2026As we cross the threshold into a new year, I am upholding my annual tradition of experimenting with the structure of The Sequence.

While SoftBank attempts to conquer AGI through capital dominance, Chinese lab DeepSeek continues to demonstrate that smarter math can rival larger budgets.

This approach enables the creation of scalable, persistent, and controllable environments—such as infinite travel atlases or galaxy simulations—that bridge the gap between static web frameworks and fully gener…

1 week, 3 days назад @ thesequence.substack.com
The Sequence Opinion #782: The New Gradient: Research Directions That Will Ship in 2026
The Sequence Opinion #782: The New Gradient: Research Directions That Will Ship in 2026 The Sequence Opinion #782: The New Gradient: Research Directions That Will Ship in 2026

As the first post of 2026, I wanted to share some of the research trends that I think might be super influential in this year frontier model breakthroughs.

We built these incredible “stochastic parrots” that could complete your code, write poetry, and pass the bar exam, all by just really, really wanting to predict the next token.

But if you look at the research papers dropping recently, the vibe has shifted.

We don’t just want models that can talk smooth; we want models that can think straight.

As we look toward 2026, we are moving from the era of Generative AI (making things that look real) to Verifiable AI (making things that are correct).

1 week, 6 days назад @ thesequence.substack.com
The Sequence AI of the Week #781: The Amazing GLM 4.7
The Sequence AI of the Week #781: The Amazing GLM 4.7 The Sequence AI of the Week #781: The Amazing GLM 4.7

In the frenetic closing weeks of 2025, the artificial intelligence landscape was dominated by the usual titans.

Google’s Gemini 3 Pro dazzled with its multimodal fluidity, and the rumor mill surrounding OpenAI’s GPT-5.2 reached a fever pitch.

Yet, amidst this cacophony of frontier model releases, a quieter, perhaps more consequential shift occurred on December 22.

Zhipu AI (Z.ai) released GLM-4.7, a model that signals a definitive transition in the open-weight ecosystem: a move away from “conversational competency” toward “agentic reliability.”GLM-4.7 is not designed to be a chat partner; it is designed to be an employee.

By aggressively optimizing for long-context loops, terminal error rec…

2 weeks назад @ thesequence.substack.com
The Sequence Knowledge # 780: Synthetic Data for Image Models
The Sequence Knowledge # 780: Synthetic Data for Image Models The Sequence Knowledge # 780: Synthetic Data for Image Models

Today we will Discuss:Key concepts of synthetic data generation for image models.

💡 AI Concept of the Day: Synthetic Data Generation for Image ModelsSynthetic image data has moved from a niche trick to a core ingredient in modern vision systems.

When real images are scarce, private, or unbalanced, synthetic pipelines let you generate pixels with known labels, push coverage into rare and long-tail cases, and iterate quickly on edge conditions.

The key is choosing the right generator, the right control signals, and a rigorous quality-control loop so that synthetic variety actually translates into downstream gains.

Diffusion models (text-to-image, image-to-image, inpainting) and GANs can produ…

2 weeks, 1 day назад @ thesequence.substack.com
The Sequence Radar #779: The Inference Wars and China’s AI IPO Race
The Sequence Radar #779: The Inference Wars and China’s AI IPO Race The Sequence Radar #779: The Inference Wars and China’s AI IPO Race

In the AI of the Week section we are going to cover the newly released GLM 4.7.

In the opinion section, we will discuss three non-trivial AI areas that I am excited about fro 2026.

For Nvidia, this isn’t just about removing a rival; it’s about acknowledging that the next phase of the AI cycle is inference.

AI Lab: Princeton University, ByteDance Seed, Columbia University, University of Michigan, University of ChicagoSummary: The authors propose GenEnv, a framework that trains agents via a co-evolutionary game where a generative simulator creates dynamic tasks aligned with the agent’s current difficulty level.

🤖 AI Tech ReleasesGLM 4.7Z.ai released GLM 4.7, a new version of its marquee model…

2 weeks, 3 days назад @ thesequence.substack.com
The Sequence Opinion #778: After Scaling: The Era of Research and New Recipes for Frontier AI
The Sequence Opinion #778: After Scaling: The Era of Research and New Recipes for Frontier AI The Sequence Opinion #778: After Scaling: The Era of Research and New Recipes for Frontier AI

Created Using GPT-5For the last few years, AI progress has felt almost… procedural.

That mood is what Ilya Sutskever calls the “age of scaling”—a period where one word (“scaling”) basically told an entire industry what to do next.

What are the techniques that convert compute into genuine generalization—models that learn faster, adapt better, and make fewer weird mistakes?

Below is a map of the most promising technique-clusters that could plausibly unlock the next wave of frontier innovation.

It’s not a list of “one weird trick.” It’s more like a toolbox for the post-pretraining world.

2 weeks, 6 days назад @ thesequence.substack.com
The Sequence AI of the Week #777: Thinking Fast, Thinking Cheap: Thinking Fast, Thinking Cheap: The Nemotron 3 Blueprint
The Sequence AI of the Week #777: Thinking Fast, Thinking Cheap: Thinking Fast, Thinking Cheap: The Nemotron 3 Blueprint The Sequence AI of the Week #777: Thinking Fast, Thinking Cheap: Thinking Fast, Thinking Cheap: The Nemotron 3 Blueprint

Last week, NVIDIA quietly redefined the baseline for open-weight intelligence with the release of the Nemotron 3 family.

The headline isn’t just the benchmarks (though the Nemotron 3 Nano creates a new state-of-the-art for the 30B parameter class); it is the architecture.

We are seeing the first major industrial-scale deployment of a Hybrid Mamba-Transformer Mixture-of-Experts (MoE).

Image Credit: NVIDIAFor the enterprise, this is a signal that the “brute force” era of dense Transformers is ending.

The Architecture: A Hybrid “Frankenstein” (in the best way)

3 weeks назад @ thesequence.substack.com
The Sequence Knwoledge #776: Fake It 'Til You Make It: How RL is Perfecting Synthetic Data.
The Sequence Knwoledge #776: Fake It 'Til You Make It: How RL is Perfecting Synthetic Data. The Sequence Knwoledge #776: Fake It 'Til You Make It: How RL is Perfecting Synthetic Data.

Created Using Gemini 3Today we will Discuss:The idea of using reinforcement learning(RL) environments to generate synthetic data.

The famous Reflexion paper about improving AI agents using RL data generation.

💡 AI Concept of the Day: Synthetic Data Generation with RL EnvironmentsWhen real-world data is scarce or privacy-restricted, reinforcement learning (RL) environments become a force multiplier for synthetic data.

This is especially potent for domains where outcomes are verifiable but logs are limited (coding sandboxes, web automation, spreadsheets/SQL, robotics-in-sim).

By executing tasks rather than describing them, RL pipelines mint trajectories that teach models how to act under cons…

3 weeks, 1 day назад @ thesequence.substack.com
The Sequence Radar #775: Last Week in AI: Tokens, Throughput, and Trillions
The Sequence Radar #775: Last Week in AI: Tokens, Throughput, and Trillions The Sequence Radar #775: Last Week in AI: Tokens, Throughput, and Trillions

In the AI of the week edition, we discuss NVIDIA’s amazing Nemotron 3 release.

Subscribe and don’t miss out:📝 Editorial: Last Week in AI: Tokens, Throughput, and TrillionsThis week’s AI story didn’t arrive as one dramatic demo; it arrived as a synchronized upgrade across the stack—capital, platforms, and product surfaces all moving in lockstep.

AI Lab: Tongyi Lab (Alibaba Group)Summary: QwenLong-L1.5 proposes an end-to-end post-training recipe for long-context reasoning, combining a scalable synthesis pipeline for multi-hop, globally-grounded tasks with stabilized long-context RL (including task-balanced sampling, task-specific advantage estimation, and AEPO).

AI Lab: NVIDIASummary: This wh…

3 weeks, 3 days назад @ thesequence.substack.com
📓 Cool Blogs
ODS.ai Habr ODS.ai Habr
последний пост 3 months, 4 weeks назад
SWE-MERA — новый динамический бенчмарк для моделей агентной генерации кода
SWE-MERA — новый динамический бенчмарк для моделей агентной генерации кода SWE-MERA — новый динамический бенчмарк для моделей агентной генерации кода

Однако все задачи в MERA CODE, как впрочем и в SWE-bench и других бенчмарках подобного назначения, следуют классической парадигме, когда у нас есть фиксированный обучающий набор данных и, что более важно, фиксированный проверочный набор.

Но большие языковые модели для кодинга, которые мы и пытаемся оценивать нашим набором, также учатся на GitHub – со времен еще первой модели LLaMa.

Кажется, что 700 задач немного, но это уже очень приличное количество, и что самое важное — это новые задачи.

Current behavior: from sympy import ask, Q, Symbol x = Symbol('x') print(ask(Q.finite(x**-1), Q.real(x))) # Output: True Expected behavior: The function should return None to indicate uncertainty, as x**-…

3 months, 4 weeks назад @ habr.com
DRAGON: динамический бенчмарк для оценки RAG-систем на русском языке
DRAGON: динамический бенчмарк для оценки RAG-систем на русском языке DRAGON: динамический бенчмарк для оценки RAG-систем на русском языке

Ответ: Кэисукэ ТибаSPARQL-запрос SimpleSELECT DISTINCT ?s ?r ?o WHERE { { SELECT ?s ?r ?o WHERE { ?s ?r ?o . }

GROUP BY ?s ?r HAVING(count(?o) = 1) } { SELECT ?s ?r ?o WHERE { ?s ?r ?o . }

Ответ: Национальная система платежных карт (НСПК) Центр биометрических технологий (ЦБТ) ЕБСSELECT ?s ?r ?o ?len WHERE { { SELECT ?s ?r (COUNT(?o1) as ?len) (GROUP_CONCAT(DISTINCT(STR(?o1));separator="|") AS ?o) WHERE { ?s ?r ?o1 . }

FILTER(?o != ?o1) } GROUP BY ?o ?o1 ?r ?r1 HAVING(COUNT(?s) = 1) } UNION { SELECT ?s ?r ?o ?r1 ?s1 WHERE { ?s ?r ?o .

FILTER(?o != ?o1) } GROUP BY ?o ?o1 ?r ?r1 HAVING(COUNT(?s) = 1) } UNION { SELECT ?s ?r ?o ?r1 ?s1 WHERE { ?s ?r ?o .

5 months, 3 weeks назад @ habr.com
RKNN Toolkit2: конвертация моделей и симуляция NPU Rockchip
RKNN Toolkit2: конвертация моделей и симуляция NPU Rockchip RKNN Toolkit2: конвертация моделей и симуляция NPU Rockchip

В этой статье я хочу поделиться своим опытом по конвертации нейросети в формат rknn с помощью библиотеки rknn-toolkit2.

Вот как выглядят веса pytorch модели в Netron:веса pytorch модели в NetronВажно!

Конвертация onnx модели в rknnДалее создается объект RKNN , который управляет процессом конвертации и инференса модели на платформе Rockchip.

На этом этапе происходит подготавка модели к конвертации в формат RKNN и последующему запуску на NPU Rockchip.

Создание и экспорт rknn моделиНа этом этапе происходит конвертация ONNX-модели во внутренний формат RKNN, оптимизация графа и подготовка к запуску на NPU Rockchip.

6 months назад @ habr.com
MERA Code: всесторонняя оценка генерации кода в прикладных сценариях
MERA Code: всесторонняя оценка генерации кода в прикладных сценариях MERA Code: всесторонняя оценка генерации кода в прикладных сценариях

🔗MERA Code🔗GitHub с кодом и данными🔗Коллекция на Hugging Face🔗Статья на arxiv🔗Репозиторий проекта на GitVerseЧто такое MERA Code?

Современные кодовые языковые модели и модели общего назначения (ChatGPT, Claude, Qwen, YandexGPT, GigaChat и др.)

Список текущих задач MERA Code и их характеристикКаталог задач MERA Code и их подробное описание представлено на сайте.

В MERA Code промпты строго подобраны под задачу и корректный выбор ответа.

В заключениеMERA Code — это попытка закрыть важный пробел в тестировании LLM: насколько они действительно полезны в реальной, локализованной разработке.

6 months назад @ habr.com
Байесовская собака: анализ пёсьего компаса
Байесовская собака: анализ пёсьего компаса Байесовская собака: анализ пёсьего компаса

", подумал я. И, к счастью, у меня как раз под рукой оказался идеальный подопытный.

Стандартное арифметическое среднее между 360° и 0° даст нам 180°, несмотря на то, что и 360°, и 0° указывают в одном направлении.

Нулевая гипотеза утверждает, что данные распределены равномерно по кругу, альтернативная — что это не так.

from pingouin import circ_vtest v, pval = circ_vtest(data['radians'], dir=np.pi) print(f"V-statistics: {v:.3f}; p-value: {pval:.6f}")>> V-statistics: 24.127; p-value: 0.002904Вот мы и подобрались к чему-то интересному!

Априорное распределение и функция правдоподобияПредположим, что у нас есть:Априорное распределение с параметрамиФункция правдоподобия для нового наблюдения с п…

9 months, 3 weeks назад @ habr.com
Machine Learning Mastery
последний пост 2 weeks, 5 days назад
Training a Model on Multiple GPUs with Data Parallelism
Training a Model on Multiple GPUs with Data Parallelism Training a Model on Multiple GPUs with Data Parallelism

device ( f "cuda:{local_rank}" ) print ( f "World size: {world_size}, Rank: {rank}, Local rank: {local_rank}.

save ( { "model" : model .

mlp = LlamaMLP ( config ) def forward ( self , hidden_states : Tensor , rope : RotaryPositionEncoding , attn_mask : Tensor ) -> Tensor : # First residual block: Self-attention residual = hidden_states hidden_states = self .

device ( f "cuda:{local_rank}" ) print ( f "World size: {world_size}, Rank: {rank}, Local rank: {local_rank}.

save ( { "model" : model .

2 weeks, 5 days назад @ machinelearningmastery.com
Train a Model Faster with torch.compile and Gradient Accumulation
Train a Model Faster with torch.compile and Gradient Accumulation Train a Model Faster with torch.compile and Gradient Accumulation

It is not the same model object you created using nn.Module , but it shares the same tensors with the original model.

This is because the compiled model is an object that shares the same weights as the original model.

state_dict ( ) , "model.pth" )The original model can be accessed from the compiled model using model._orig_mod .

This can be achieved by increasing the batch size: with the same number of data samples, a larger batch size means fewer batches to process.

You also learned that gradient accumulation is a technique for training with a larger effective batch size by accumulating gradients from multiple mini-batches.

2 weeks, 6 days назад @ machinelearningmastery.com
Training a Model with Limited Memory using Mixed Precision and Gradient Checkpointing
Training a Model with Limited Memory using Mixed Precision and Gradient Checkpointing Training a Model with Limited Memory using Mixed Precision and Gradient Checkpointing

set_postfix ( loss = loss .

save ( { "model" : model .

state_dict ( ) , "optimizer" : optimizer .

state_dict ( ) , "scaler" : scaler .

If you still encounter memory issues, another technique trades time for memory: gradient checkpointing.

3 weeks назад @ machinelearningmastery.com
Practical Agentic Coding with Google Jules
Practical Agentic Coding with Google Jules Practical Agentic Coding with Google Jules

Share Post ShareIntroducing Google JulesIf you have an interest in agentic coding, there’s a pretty good chance you’ve heard of Google Jules by now.

Using Google Jules with an Existing GitHub RepositoryThe process for using Google Jules generally involves initial setup, task prompting, and subsequent human review and approval.

Step 1: Initial Access and AuthenticationVisit the official Google Jules website and click “Try Jules”.

These are just a few specific concerns to keep in mind while exploring Google Jules and agentic coding.

Wrapping UpThis has been an introduction to using Google Jules for agentic coding.

3 weeks назад @ machinelearningmastery.com
Evaluating Perplexity on Language Models
Evaluating Perplexity on Language Models Evaluating Perplexity on Language Models

When you train a language model, you want to measure how accurately it predicts human language use.

Specifically, you will learn:What is perplexity, and how to compute itHow to evaluate the perplexity of a language model with sample dataLet’s get started.

OverviewThis article is divided into two parts; they are:What Is Perplexity and How to Compute ItEvaluate the Perplexity of a Language Model with HellaSwag DatasetWhat Is Perplexity and How to Compute ItPerplexity is a measure of how well a language model predicts a sample of text.

Evaluate the Perplexity of a Language Model with HellaSwag DatasetPerplexity is a dataset-dependent metric.

from_pretrained ( model ) model = transformers .

3 weeks, 1 day назад @ machinelearningmastery.com
3 Smart Ways to Encode Categorical Features for Machine Learning
3 Smart Ways to Encode Categorical Features for Machine Learning 3 Smart Ways to Encode Categorical Features for Machine Learning

Share Post ShareIn this article, you will learn three reliable techniques — ordinal encoding, one-hot encoding, and target (mean) encoding — for turning categorical features into model-ready numbers while preserving their meaning.

Applying target (mean) encoding for high-cardinality features without leaking the target.

fit_transform ( data ) print ( "Original Data:" , data .

The answer lies in Target Encoding, also frequently called Mean Encoding.

Target (Mean) Encoding: This is the powerful answer for high-cardinality features that would overwhelm OHE.

3 weeks, 2 days назад @ machinelearningmastery.com
Pretraining a Llama Model on Your Local GPU
Pretraining a Llama Model on Your Local GPU Pretraining a Llama Model on Your Local GPU

Once trained, you can load it back with the following code:from tokenizers import Tokenizer tokenizer = Tokenizer.from_file("bpe_50k.json") 1 2 3 from tokenizers import Tokenizer tokenizer = Tokenizer .

tokenizer = tokenizer self .

device = device self .

save ( { "model" : model .

save ( { "model" : model .

3 weeks, 2 days назад @ machinelearningmastery.com
Rotary Position Embeddings for Long Context Length
Rotary Position Embeddings for Long Context Length Rotary Position Embeddings for Long Context Length

dim = dim self .

cat ( ( inv_freq , inv_freq ) , dim = - 1 ) position = torch .

dim = dim self .

pi / inv_freq max_wavelen = base_length / low_factor min_wavelen = base_length / high_factor smooth_factor = ( base_length / wavelen - low_factor ) / ( high_factor - low_factor ) smoothed = ( 1 - smooth_factor ) * inv_freq / scale_factor + smooth_factor * inv_freq inv_freq = torch .

cat ( ( inv_freq , inv_freq ) , dim = - 1 ) position = torch .

3 weeks, 4 days назад @ machinelearningmastery.com
How to Fine-Tune a Local Mistral or Llama 3 Model on Your Own Dataset
How to Fine-Tune a Local Mistral or Llama 3 Model on Your Own Dataset How to Fine-Tune a Local Mistral or Llama 3 Model on Your Own Dataset

In this tutorial, we’ll learn how to fine-tune two powerful open-source models, Mistral 7B and Llama 3 8B, using a customer support question-and-answer dataset.

print("Creating customer support Q&A dataset...") # Create realistic customer support data customer_support_data = [ { "instruction": "You are a helpful customer support agent.

decode ( outputs [ 0 ] , skip_special_tokens = True ) # Extract just the response text if "[/INST]" in response : response = response .

strip ( ) elif "assistant" in response : response = response .

strip ( ) elif "### Response:" in response : response = response .

3 weeks, 5 days назад @ machinelearningmastery.com
5 Agentic Coding Tips & Tricks
5 Agentic Coding Tips & Tricks 5 Agentic Coding Tips & Tricks

A practical trick is a diff budget, an explicit limit on lines changed per iteration.

startswith ( ( "+++" , "---" ) ) ) changed = count_changed_lines ( diff ) if changed > MAX_CHANGED_LINES : raise ValueError ( f "Diff too large: {changed} changed lines" )For manual workflows, bake the constraint into your prompt:Output only a unified diffHard limit: 120 changed lines totalNo unrelated formatting or refactorsIf you need more, stop and ask for a second patchAgents respond well to constraints that are measurable.

ratelimit import SlidingWindowLimiter def test_allows_n_requests_per_window ( ) : lim = SlidingWindowLimiter ( limit = 3 , window_seconds = 1 ) assert lim .

allow ( "u1" ) assert li…

3 weeks, 6 days назад @ machinelearningmastery.com
The Real Cost of Inaction: How Silos Hurt Productivity for Data Scientists (Sponsored)
The Real Cost of Inaction: How Silos Hurt Productivity for Data Scientists (Sponsored) The Real Cost of Inaction: How Silos Hurt Productivity for Data Scientists (Sponsored) 4 weeks назад @ bit.ly
Top 5 Vector Databases for High-Performance LLM Applications
Top 5 Vector Databases for High-Performance LLM Applications Top 5 Vector Databases for High-Performance LLM Applications

Vector databases solve this by storing embeddings and facilitating super-fast similarity searches across billions of vectors.

This article covers the top five vector databases for production LLM applications.

WeaviateWeaviate is an open-source vector database that works well for combining vector search with traditional database capabilities.

If you already use PostgreSQL and would like to explore a vector search extension, you can also consider pgvector.

To learn more about how vector databases work, read The Complete Guide to Vector Databases for Machine Learning.

4 weeks назад @ machinelearningmastery.com
The Machine Learning Engineer’s Checklist: Best Practices for Reliable Models
The Machine Learning Engineer’s Checklist: Best Practices for Reliable Models The Machine Learning Engineer’s Checklist: Best Practices for Reliable Models

Share Post ShareIntroductionBuilding newly trained machine learning models that work is a relatively straightforward endeavor, thanks to mature frameworks and accessible computing power.

Without further ado, here is the list of 10 machine learning engineer best practices I curated for you and your upcoming models to shine at their best in terms of long-term reliability.

Therefore, everything surrounding a machine learning model should be properly versioned.

Continuous Monitoring and ObservabilityThis is probably already in your checklist of best practices, but as an essential of machine learning engineering, it is worth pointing it out.

This article provided a checklist of 10 essential best…

4 weeks назад @ machinelearningmastery.com
Transformer vs LSTM for Time Series: Which Works Better?
Transformer vs LSTM for Time Series: Which Works Better? Transformer vs LSTM for Time Series: Which Works Better?

IntroductionFrom daily weather measurements or traffic sensor readings to stock prices, time series data are present nearly everywhere.

lstm = nn .

transformer = nn .

Linear ( d_model , 1 ) def forward ( self , x ) : x = self .

embed ( x ) x = self .

1 month назад @ machinelearningmastery.com
How LLMs Choose Their Words: A Practical Walk-Through of Logits, Softmax and Sampling
How LLMs Choose Their Words: A Practical Walk-Through of Logits, Softmax and Sampling How LLMs Choose Their Words: A Practical Walk-Through of Logits, Softmax and Sampling

This randomness is not a bug but a core feature of how the model samples its next token from a probability distribution.

Softmax transforms these raw scores into a probability distribution.

LLMs don’t always select the token with the highest probability; instead, they sample from the probability distribution to produce a different output each time.

Top-$p$ sampling (also known as nucleus sampling) addresses this issue by sampling tokens according to their cumulative probability rather than a fixed count.

You learned to select different values for the temperature, top-$k$, and top-$p$ sampling parameters for different LLM use cases.

1 month назад @ machinelearningmastery.com
ML in Production
последний пост None
Sorta Insightful Sorta Insightful
последний пост 1 month, 4 weeks назад
Authentic Imperfection
Authentic Imperfection Authentic Imperfection

* * *I’ve been thinking about the anger surrounding generative AI.

To keep things fair, he took the best human images and best AI images, meaning human art from famous artists, and AI art from prompters skilled at removing obvious tells of image generation.

When people complain about AI slop, I see it as a complaint against the deluge of default style AI images.

We’ve seen this happen in all forms: AI text, AI music, older forms of computer generated content like CGI.

As much as we celebrate imperfection, digital imperfection is a step too far.

1 month, 4 weeks назад @ alexirpan.com
Ten Years Later
Ten Years Later Ten Years Later

Every now and then, someone asks me why I blog, and I don’t know really know what to tell them.

That’s another reason I’m not celebrating 10 years with more gusto, I know I’ve been writing less.

Indiana Jones and the Great Circle: I don’t know how they did it, but Indiana Jones and the Great Circle was just fun all the way through.

My one complaint is that the hand-to-hand combat feels like the worst part of the game, so of course they put a bunch of upgrades behind learning parry timings you’ll never use later.

I have not tried Peak, but Another Crab’s Treasure was really good and is worth playing if you’re interested in a Souls-like.

4 months, 4 weeks назад @ alexirpan.com
Brony Musicians Seize The Means of Production: My Eyewitness Account to BABSCon 2025
Brony Musicians Seize The Means of Production: My Eyewitness Account to BABSCon 2025 Brony Musicians Seize The Means of Production: My Eyewitness Account to BABSCon 2025

A music concert in the evenings, typically set up as a rave with EDM or rock music made by brony musicians.

She has been involved in organizing pony music concerts for over a decade, for both BABSCon and other pony conventions.

Thank you, BABSCon ChairsThe brony musicians immediately jump into an emergency Discord call with Pinkaboo, to get her side of the story.

Other conventions start tweeting in support of the brony musicians, with no one taking BABSCon’s side.

It’s hard for me to explain why I like MLP fan music, because brony music really isn’t accessible.

5 months, 3 weeks назад @ alexirpan.com
Who is AI For?
Who is AI For? Who is AI For?

I think the easy answer to this question is that right now, AI is for the AI developers.

Code is useful, it makes money, it is a testbed for AI speeding up the development of AI, and it is easy.

I’m working in AI because it pays well and is potentially really good for the world.

The artists did not know what AI was, but when they learned, they quickly decided they did not want it.

It feels like the most likely outcome is that people go all-in on pushing raw intelligence, in the way that AI developers can measure it, leaving behind those that are not like AI developers.

9 months, 2 weeks назад @ alexirpan.com
Lil'Log
последний пост None
The Spectator
последний пост None
Off the Convex Path
последний пост None
Piekniewski's blog
последний пост None
fast.ai NLP fast.ai NLP
последний пост None
Sebastian Ruder
последний пост None
Andrew Karpathy blog
последний пост None
大トロ 大トロ
последний пост None
🔬 Science
Papers With Code Papers With Code
последний пост 5 months, 3 weeks назад
/henry123-boy/ SpatialTrackerV2: 3D Point Tracking Made Easy
/henry123-boy/ SpatialTrackerV2: 3D Point Tracking Made Easy /henry123-boy/ SpatialTrackerV2: 3D Point Tracking Made Easy

We present SpatialTrackerV2, a feed-forward 3D point tracking method for monocular videos.

Going beyond modular pipelines built on off-the-shelf components for 3D tracking, our approach unifies the intrinsic connections between point tracking, monocular depth, and camera pose estimation into a high-performing and feedforward 3D point tracker.

It decomposes world-space 3D motion into scene geometry, camera ego-motion, and pixel-wise object motion, with a fully differentiable and end-to-end architecture, allowing scalable training across a wide range of datasets, including synthetic sequences, posed RGB-D videos, and unlabeled in-the-wild footage.

By learning geometry and motion jointly from …

5 months, 3 weeks назад @ paperswithcode.com
/antof27/ Efficient Calisthenics Skills Classification through Foreground Instance Selection and Depth Estimation
/antof27/ Efficient Calisthenics Skills Classification through Foreground Instance Selection and Depth Estimation /antof27/ Efficient Calisthenics Skills Classification through Foreground Instance Selection and Depth Estimation

Calisthenics skill classification is the computer vision task of inferring the skill performed by an athlete from images, enabling automatic performance assessment and personalized analytics.

Traditional methods for calisthenics skill recognition are based on pose estimation methods to determine the position of skeletal data from images, which is later fed to a classification algorithm to infer the performed skill.

This work proposes a direct approach to calisthenics skill recognition, which leverages depth estimation and athlete patch retrieval to avoid the computationally expensive human pose estimation module.

Using Depth Anything V2 for depth estimation and YOLOv10 for athlete localizat…

5 months, 3 weeks назад @ paperswithcode.com
/snowflakedb/ Arctic Inference with Shift Parallelism: Fast and Efficient Open Source Inference System for Enterprise AI
/snowflakedb/ Arctic Inference with Shift Parallelism: Fast and Efficient Open Source Inference System for Enterprise AI /snowflakedb/ Arctic Inference with Shift Parallelism: Fast and Efficient Open Source Inference System for Enterprise AI

Inference is now the dominant AI workload, yet existing systems force trade-offs between latency, throughput, and cost.

Arctic Inference, an open-source vLLM plugin from Snowflake AI Research, introduces Shift Parallelism, a dynamic parallelism strategy that adapts to real-world traffic while integrating speculative decoding, SwiftKV compute reduction, and optimized embedding inference.

It achieves up to 3.4 times faster request completion, 1.75 times faster generation, and 1.6M tokens/sec per GPU for embeddings, outperforming both latency- and throughput-optimized deployments.

Already powering Snowflake Cortex AI, Arctic Inference delivers state-of-the-art, cost-effective inference for ent…

5 months, 3 weeks назад @ paperswithcode.com
/NVIDIA/ FourCastNet 3: A geometric approach to probabilistic machine-learning weather forecasting at scale
/NVIDIA/ FourCastNet 3: A geometric approach to probabilistic machine-learning weather forecasting at scale /NVIDIA/ FourCastNet 3: A geometric approach to probabilistic machine-learning weather forecasting at scale

FourCastNet 3 advances global weather modeling by implementing a scalable, geometric machine learning (ML) approach to probabilistic ensemble forecasting.

The approach is designed to respect spherical geometry and to accurately model the spatially correlated probabilistic nature of the problem, resulting in stable spectra and realistic dynamics across multiple scales.

FourCastNet 3 delivers forecasting accuracy that surpasses leading conventional ensemble models and rivals the best diffusion-based methods, while producing forecasts 8 to 60 times faster than these approaches.

In contrast to other ML approaches, FourCastNet 3 demonstrates excellent probabilistic calibration and retains realis…

5 months, 3 weeks назад @ paperswithcode.com
/jingyanw/ Choosing the Better Bandit Algorithm under Data Sharing: When Do A/B Experiments Work?
/jingyanw/ Choosing the Better Bandit Algorithm under Data Sharing: When Do A/B Experiments Work? /jingyanw/ Choosing the Better Bandit Algorithm under Data Sharing: When Do A/B Experiments Work?

We study A/B experiments that are designed to compare the performance of two recommendation algorithms.

The bias arising from this type of data sharing is known as "symbiosis bias".

In this paper, we highlight that, for decision-making purposes, the sign of the GTE often matters more than its precise magnitude when selecting the better algorithm.

We formalize this insight under a multi-armed bandit framework and theoretically characterize when the sign of the expected GTE estimate under data sharing aligns with or contradicts the sign of the true GTE.

Our analysis identifies the level of exploration versus exploitation as a key determinant of how symbiosis bias impacts algorithm selection.

5 months, 3 weeks назад @ paperswithcode.com
/qqq-yi/ DAC: A Dynamic Attention-aware Approach for Task-Agnostic Prompt Compression
/qqq-yi/ DAC: A Dynamic Attention-aware Approach for Task-Agnostic Prompt Compression /qqq-yi/ DAC: A Dynamic Attention-aware Approach for Task-Agnostic Prompt Compression

Task-agnostic prompt compression leverages the redundancy in natural language to reduce computational overhead and enhance information density within prompts, especially in long-context scenarios.

Existing methods predominantly rely on information entropy as the metric to compress lexical units, aiming to achieve minimal information loss.

However, these approaches overlook two critical aspects: (i) the importance of attention-critical tokens at the algorithmic level, and (ii) shifts in information entropy during the compression process.

Motivated by these challenges, we propose a dynamic attention-aware approach for task-agnostic prompt compression (DAC).

This approach effectively integrate…

5 months, 3 weeks назад @ paperswithcode.com
/lukasellinger/ Simplifications are Absolutists: How Simplified Language Reduces Word Sense Awareness in LLM-Generated Definitions
/lukasellinger/ Simplifications are Absolutists: How Simplified Language Reduces Word Sense Awareness in LLM-Generated Definitions /lukasellinger/ Simplifications are Absolutists: How Simplified Language Reduces Word Sense Awareness in LLM-Generated Definitions

Large Language Models (LLMs) can provide accurate word definitions and explanations for any context.

However, the scope of the definition changes for different target groups, like children or language learners.

We investigate how simplification impacts homonym definition quality across three target groups: Normal, Simple, and ELI5.

Our results show that simplification drastically degrades definition completeness by neglecting polysemy, increasing the risk of misunderstanding.

Fine-tuning Llama 3.1 8B with Direct Preference Optimization substantially improves homonym response quality across all prompt types.

5 months, 3 weeks назад @ paperswithcode.com
/pspdada/ Mitigating Object Hallucinations via Sentence-Level Early Intervention
/pspdada/ Mitigating Object Hallucinations via Sentence-Level Early Intervention /pspdada/ Mitigating Object Hallucinations via Sentence-Level Early Intervention

Multimodal large language models (MLLMs) have revolutionized cross-modal understanding but continue to struggle with hallucinations - fabricated content contradicting visual inputs.

Existing hallucination mitigation methods either incur prohibitive computational costs or introduce distribution mismatches between training data and model outputs.

We identify a critical insight: hallucinations predominantly emerge at the early stages of text generation and propagate through subsequent outputs.

To address this, we propose **SENTINEL** (**S**entence-level **E**arly i**N**tervention **T**hrough **IN**-domain pr**E**ference **L**earning), a framework that eliminates dependency on human annotations…

5 months, 3 weeks назад @ paperswithcode.com
/owos/ FLEXITOKENS: Flexible Tokenization for Evolving Language Models
/owos/ FLEXITOKENS: Flexible Tokenization for Evolving Language Models /owos/ FLEXITOKENS: Flexible Tokenization for Evolving Language Models

Language models (LMs) are challenging to adapt to new data distributions by simple finetuning.

This is due to the rigidity of their subword tokenizers, which typically remain unchanged during adaptation.

This inflexibility often leads to inefficient tokenization, causing overfragmentation of out-of-distribution domains, unseen languages, or scripts.

In this work, we develop byte-level LMs with learnable tokenizers to make tokenization adaptive.

Our models include a submodule that learns to predict boundaries between the input byte sequence, encoding it into variable-length segments.

5 months, 3 weeks назад @ paperswithcode.com
/wojiufukele/ Graph-Structured Data Analysis of Component Failure in Autonomous Cargo Ships Based on Feature Fusion
/wojiufukele/ Graph-Structured Data Analysis of Component Failure in Autonomous Cargo Ships Based on Feature Fusion /wojiufukele/ Graph-Structured Data Analysis of Component Failure in Autonomous Cargo Ships Based on Feature Fusion

To address the challenges posed by cascading reactions caused by component failures in autonomous cargo ships (ACS) and the uncertainties in emergency decision-making, this paper proposes a novel hybrid feature fusion framework for constructing a graph-structured dataset of failure modes.

A hierarchical feature fusion framework is constructed, using Word2Vec encoding to encode subsystem/component features, BERT-KPCA to process failure modes/reasons, and Sentence-BERT to quantify the semantic association between failure impact and emergency decision-making.

The dataset covers 12 systems, 1,262 failure modes, and 6,150 propagation paths.

In the label prediction results, the Shore-based Meteor…

5 months, 3 weeks назад @ paperswithcode.com
/YF-W/ Tri-Learn Graph Fusion Network for Attributed Graph Clustering
/YF-W/ Tri-Learn Graph Fusion Network for Attributed Graph Clustering /YF-W/ Tri-Learn Graph Fusion Network for Attributed Graph Clustering

In recent years, models based on Graph Convolutional Networks (GCN) have made significant strides in the field of graph data analysis.

Although the Graph Transformer architecture has mitigated some of these issues, its performance is still limited when processing heterogeneous graph data.

To address these challenges, this study proposes a novel deep clustering framework that comprising GCN, Autoencoder (AE), and Graph Transformer, termed the Tri-Learn Graph Fusion Network (Tri-GFN).

The tri-learning mechanism allows mutual learning among these modules, while the feature fusion strategy enables the model to capture complex relationships, yielding highly discriminative representations for gra…

5 months, 3 weeks назад @ paperswithcode.com
/mr-ravin/ APTx Neuron: A Unified Trainable Neuron Architecture Integrating Activation and Computation
/mr-ravin/ APTx Neuron: A Unified Trainable Neuron Architecture Integrating Activation and Computation /mr-ravin/ APTx Neuron: A Unified Trainable Neuron Architecture Integrating Activation and Computation

We propose the APTx Neuron, a novel, unified neural computation unit that integrates non-linear activation and linear transformation into a single trainable expression.

The APTx Neuron is derived from the APTx activation function, thereby eliminating the need for separate activation layers and making the architecture both computationally efficient and elegant.

The proposed neuron follows the functional form $y = \sum_{i=1}^{n} ((\alpha_i + \tanh(\beta_i x_i)) \cdot \gamma_i x_i) + \delta$, where all parameters $\alpha_i$, $\beta_i$, $\gamma_i$, and $\delta$ are trainable.

We validate our APTx Neuron-based architecture on the MNIST dataset, achieving up to 96.69\% test accuracy in just 20 ep…

5 months, 3 weeks назад @ paperswithcode.com
/Rec4Fun/ A Reproducibility Study of Product-side Fairness in Bundle Recommendation
/Rec4Fun/ A Reproducibility Study of Product-side Fairness in Bundle Recommendation /Rec4Fun/ A Reproducibility Study of Product-side Fairness in Bundle Recommendation

While this problem has been widely studied in traditional recommendation settings, its implications for bundle recommendation (BR) remain largely unexplored.

Existing fairness frameworks and metrics designed for traditional recommender systems may not directly translate to this multi-layered setting.

In this paper, we conduct a comprehensive reproducibility study of product-side fairness in BR across three real-world datasets using four state-of-the-art BR methods.

We analyze exposure disparities at both the bundle and item levels using multiple fairness metrics, uncovering important patterns.

Overall, our findings offer actionable insights for building fairer bundle recommender systems and…

5 months, 3 weeks назад @ paperswithcode.com
/cbobed/ OntView: What you See is What you Meant
/cbobed/ OntView: What you See is What you Meant /cbobed/ OntView: What you See is What you Meant

However, the lack of tools that provide effective visualization is still a significant challenge.

In this paper, we present OntView, an ontology viewer that is designed to provide users with an intuitive visual representation of ontology concepts and their formal definitions through a user-friendly interface.

Building on the use of a DL reasoner, OntView follows a "What you see is what you meant" paradigm, showing the actual inferred knowledge.

One key aspect for this is its ability to visualize General Concept Inclusions (GCI), a feature absent in existing visualization tools.

OntView has been released with an open-source license for the whole community.

5 months, 3 weeks назад @ paperswithcode.com
/Rec4Fun/ RaMen: Multi-Strategy Multi-Modal Learning for Bundle Construction
/Rec4Fun/ RaMen: Multi-Strategy Multi-Modal Learning for Bundle Construction /Rec4Fun/ RaMen: Multi-Strategy Multi-Modal Learning for Bundle Construction

These approaches fail to capture elaborate relations hidden in real-world bundle structures, resulting in suboptimal bundle representations.

To overcome this limitation, we propose RaMen, a novel method that provides a holistic multi-strategy approach for bundle construction.

RaMen utilizes both intrinsic (characteristics) and extrinsic (collaborative signals) information to model bundle structures through Explicit Strategy-aware Learning (ESL) and Implicit Strategy-aware Learning (ISL).

Integrating diverse strategies enables RaMen to learn more comprehensive and robust bundle representations.

Meanwhile, Multi-strategy Alignment & Discrimination module is employed to facilitate knowledge tr…

5 months, 3 weeks назад @ paperswithcode.com
Papers With Code Papers With Code
последний пост 5 months, 3 weeks назад
/PrimisAI/ Adaptive Multi-Agent Reasoning via Automated Workflow Generation
/PrimisAI/ Adaptive Multi-Agent Reasoning via Automated Workflow Generation /PrimisAI/ Adaptive Multi-Agent Reasoning via Automated Workflow Generation

The rise of Large Reasoning Models (LRMs) promises a significant leap forward in language model capabilities, aiming to tackle increasingly sophisticated tasks with unprecedented efficiency and accuracy.

However, despite their impressive performance, recent studies have highlighted how current reasoning models frequently fail to generalize to novel, unseen problems, often resorting to memorized solutions rather than genuine inferential reasoning.

In this paper, we introduce Nexus Architect, an enhanced iteration of our multi-agent system framework, Nexus, equipped with a novel automated workflow synthesis mechanism.

Given a user's prompt and a small set of representative examples, the Archi…

5 months, 3 weeks назад @ paperswithcode.com
/sharanya02/ Real Time Captioning of Sign Language Gestures in Video Meetings
/sharanya02/ Real Time Captioning of Sign Language Gestures in Video Meetings /sharanya02/ Real Time Captioning of Sign Language Gestures in Video Meetings

One of the most tested ways to establish such a communication is through the use of sign based languages.

However, not many people are aware of the smaller intricacies involved with sign language.

Sign language recognition using computer vision aims at eliminating the communication barrier between deaf-mute and ordinary people so that they can properly communicate with others.

In recent studies, it has been found that people with hearing disabilities prefer to sign over typing during these video calls.

In this paper, we are proposing a browser extension that will automatically translate sign language to subtitles for everyone else in the video call.

5 months, 3 weeks назад @ paperswithcode.com
/alessiopittiglio/ Leveraging Context for Multimodal Fallacy Classification in Political Debates
/alessiopittiglio/ Leveraging Context for Multimodal Fallacy Classification in Political Debates /alessiopittiglio/ Leveraging Context for Multimodal Fallacy Classification in Political Debates

In this paper, we present our submission to the MM-ArgFallacy2025 shared task, which aims to advance research in multimodal argument mining, focusing on logical fallacies in political debates.

Our approach uses pretrained Transformer-based models and proposes several ways to leverage context.

In the fallacy classification subtask, our models achieved macro F1-scores of 0.4444 (text), 0.3559 (audio), and 0.4403 (multimodal).

Our multimodal model showed performance comparable to the text-only model, suggesting potential for improvements.

PDFAbstract

5 months, 3 weeks назад @ paperswithcode.com
/RS2002/ One Step is Enough: Multi-Agent Reinforcement Learning based on One-Step Policy Optimization for Order Dispatch on Ride-Sharing Platforms
/RS2002/ One Step is Enough: Multi-Agent Reinforcement Learning based on One-Step Policy Optimization for Order Dispatch on Ride-Sharing Platforms /RS2002/ One Step is Enough: Multi-Agent Reinforcement Learning based on One-Step Policy Optimization for Order Dispatch on Ride-Sharing Platforms

On-demand ride-sharing platforms face the fundamental challenge of dynamically bundling passengers with diverse origins and destinations and matching them with vehicles in real time, all under significant uncertainty.

However, conventional MARL-based ride-sharing approaches heavily rely on the accurate estimation of Q-values or V-values, which becomes problematic in large-scale, highly uncertain environments.

To address these challenges, we propose two novel alternative methods that bypass value function estimation.

First, we adapt GRPO to ride-sharing, replacing the PPO baseline with the group average reward to eliminate critic estimation errors and reduce training bias.

Second, inspired b…

5 months, 3 weeks назад @ paperswithcode.com
/LiXinran6/ Long-Short Distance Graph Neural Networks and Improved Curriculum Learning for Emotion Recognition in Conversation
/LiXinran6/ Long-Short Distance Graph Neural Networks and Improved Curriculum Learning for Emotion Recognition in Conversation /LiXinran6/ Long-Short Distance Graph Neural Networks and Improved Curriculum Learning for Emotion Recognition in Conversation

Include the markdown at the top of your GitHub README.md file to showcase the performance of the model.

Badges are live and will be dynamically updated with the latest ranking of this paper.

5 months, 3 weeks назад @ paperswithcode.com
/ShimSoonYong/ ZClassifier: Temperature Tuning and Manifold Approximation via KL Divergence on Logit Space
/ShimSoonYong/ ZClassifier: Temperature Tuning and Manifold Approximation via KL Divergence on Logit Space

We introduce a novel classification framework, ZClassifier, that replaces conventional deterministic logits with diagonal Gaussian-distributed logits. Code: https://github.com/ShimSoonYong/ZClassifier

5 months, 4 weeks назад @ paperswithcode.com
/briziorusso/ On Gradual Semantics for Assumption-Based Argumentation
/briziorusso/ On Gradual Semantics for Assumption-Based Argumentation

In this paper, we fill this gap and propose a family of novel gradual semantics for equipping assumptions, which are the core components in ABA frameworks, with dialectical strengths. Code: https://github.com/briziorusso/GradualABA

5 months, 4 weeks назад @ paperswithcode.com
/wumingqi/ Reasoning or Memorization? Unreliable Results of Reinforcement Learning Due to Data Contamination
/wumingqi/ Reasoning or Memorization? Unreliable Results of Reinforcement Learning Due to Data Contamination

Cloudflare is unable to establish an SSL connection to the origin server.

If you're a visitor of this website:Please try again in a few minutes.

If you're the owner of this website:It appears that the SSL configuration used is not compatible with Cloudflare.

This could happen for a several reasons, including no shared cipher suites.

Additional troubleshooting information here.

5 months, 4 weeks назад @ paperswithcode.com
/IsaacYQH/ WildFX: A DAW-Powered Pipeline for In-the-Wild Audio FX Graph Modeling
/IsaacYQH/ WildFX: A DAW-Powered Pipeline for In-the-Wild Audio FX Graph Modeling

Despite rapid progress in end-to-end AI music generation, AI-driven modeling of professional Digital Signal Processing (DSP) workflows remains challenging. Code: https://github.com/IsaacYQH/WildFX

5 months, 4 weeks назад @ paperswithcode.com
/summer1278/ Addressing Data Imbalance in Transformer-Based Multi-Label Emotion Detection with Weighted Loss
/summer1278/ Addressing Data Imbalance in Transformer-Based Multi-Label Emotion Detection with Weighted Loss

This paper explores the application of a simple weighted loss function to Transformer-based models for multi-label emotion detection in SemEval-2025 Shared Task 11. Code: https://github.com/summer1278/semeval2025-task11

5 months, 4 weeks назад @ paperswithcode.com
/gabrielkmbo/ Step-wise Policy for Rare-tool Knowledge (SPaRK): Offline RL that Drives Diverse Tool Use in LLMs
/gabrielkmbo/ Step-wise Policy for Rare-tool Knowledge (SPaRK): Offline RL that Drives Diverse Tool Use in LLMs

We present Step-wise Policy for Rare-tool Knowledge (SPaRK), a novel reinforcement learning framework that teaches large language models to explore diverse tool usage patterns beyond conventional high-temperature sampling. Code: https://github.com/gabrielkmbo/explore-rl

5 months, 4 weeks назад @ paperswithcode.com
/Cavendish518/ Learning to Tune Like an Expert: Interpretable and Scene-Aware Navigation via MLLM Reasoning and CVAE-Based Adaptation
/Cavendish518/ Learning to Tune Like an Expert: Interpretable and Scene-Aware Navigation via MLLM Reasoning and CVAE-Based Adaptation

Service robots are increasingly deployed in diverse and dynamic environments, where both physical layouts and social contexts change over time and across locations. Code: https://github.com/Cavendish518/LE-Nav

5 months, 4 weeks назад @ paperswithcode.com
/MatteoFasulo/ AI Wizards at CheckThat! 2025: Enhancing Transformer-Based Embeddings with Sentiment for Subjectivity Detection in News Articles
/MatteoFasulo/ AI Wizards at CheckThat! 2025: Enhancing Transformer-Based Embeddings with Sentiment for Subjectivity Detection in News Articles

Cloudflare is unable to establish an SSL connection to the origin server.

If you're a visitor of this website:Please try again in a few minutes.

If you're the owner of this website:It appears that the SSL configuration used is not compatible with Cloudflare.

This could happen for a several reasons, including no shared cipher suites.

Additional troubleshooting information here.

5 months, 4 weeks назад @ paperswithcode.com
/VCA-EPFL/ SystolicAttention: Fusing FlashAttention within a Single Systolic Array
/VCA-EPFL/ SystolicAttention: Fusing FlashAttention within a Single Systolic Array

The frequent data swaps between the systolic array and external vector units result in low systolic array utilization. Code: https://github.com/VCA-EPFL/FSA

5 months, 4 weeks назад @ paperswithcode.com
/Buddhi19/ Precision Spatio-Temporal Feature Fusion for Robust Remote Sensing Change Detection
/Buddhi19/ Precision Spatio-Temporal Feature Fusion for Robust Remote Sensing Change Detection

Cloudflare is unable to establish an SSL connection to the origin server.

If you're a visitor of this website:Please try again in a few minutes.

If you're the owner of this website:It appears that the SSL configuration used is not compatible with Cloudflare.

This could happen for a several reasons, including no shared cipher suites.

Additional troubleshooting information here.

5 months, 4 weeks назад @ paperswithcode.com
Papers With Code Papers With Code
последний пост 5 months, 3 weeks назад
/fudanvi/ Beyond Task-Specific Reasoning: A Unified Conditional Generative Framework for Abstract Visual Reasoning
/fudanvi/ Beyond Task-Specific Reasoning: A Unified Conditional Generative Framework for Abstract Visual Reasoning

Cloudflare is unable to establish an SSL connection to the origin server.

If you're a visitor of this website:Please try again in a few minutes.

If you're the owner of this website:It appears that the SSL configuration used is not compatible with Cloudflare.

This could happen for a several reasons, including no shared cipher suites.

Additional troubleshooting information here.

5 months, 4 weeks назад @ paperswithcode.com
/benedekrozemberczki/ PGT-I: Scaling Spatiotemporal GNNs with Memory-Efficient Distributed Training
/benedekrozemberczki/ PGT-I: Scaling Spatiotemporal GNNs with Memory-Efficient Distributed Training

Spatiotemporal graph neural networks (ST-GNNs) are powerful tools for modeling spatial and temporal data dependencies. Code: https://github.com/benedekrozemberczki/pytorch_geometric_temporal

5 months, 4 weeks назад @ paperswithcode.com
/chengxuphd/ DCR: Quantifying Data Contamination in LLMs Evaluation
/chengxuphd/ DCR: Quantifying Data Contamination in LLMs Evaluation

Cloudflare is unable to establish an SSL connection to the origin server.

If you're a visitor of this website:Please try again in a few minutes.

If you're the owner of this website:It appears that the SSL configuration used is not compatible with Cloudflare.

This could happen for a several reasons, including no shared cipher suites.

Additional troubleshooting information here.

5 months, 4 weeks назад @ paperswithcode.com
/gitter-lab/ Assay2Mol: large language model-based drug design using BioAssay context
/gitter-lab/ Assay2Mol: large language model-based drug design using BioAssay context

Scientific databases aggregate vast amounts of quantitative data alongside descriptive text. Code: https://github.com/gitter-lab/Assay2Mol

5 months, 4 weeks назад @ paperswithcode.com
/hayatkhan8660-maker/ DVFL-Net: A Lightweight Distilled Video Focal Modulation Network for Spatio-Temporal Action Recognition
/hayatkhan8660-maker/ DVFL-Net: A Lightweight Distilled Video Focal Modulation Network for Spatio-Temporal Action Recognition

We employ forward Kullback-Leibler (KL) divergence alongside spatio-temporal focal modulation to effectively transfer both local and global context from the Video-FocalNet Base (teacher) to the proposed VFL-Net (student). Code: https://github.com/hayatkhan8660-maker/DVFL-Net

5 months, 4 weeks назад @ paperswithcode.com
/JudyJuezhuLong/ Best Practices for Large-Scale, Pixel-Wise Crop Mapping and Transfer Learning Workflows
/JudyJuezhuLong/ Best Practices for Large-Scale, Pixel-Wise Crop Mapping and Transfer Learning Workflows

Cloudflare is unable to establish an SSL connection to the origin server.

If you're a visitor of this website:Please try again in a few minutes.

If you're the owner of this website:It appears that the SSL configuration used is not compatible with Cloudflare.

This could happen for a several reasons, including no shared cipher suites.

Additional troubleshooting information here.

5 months, 4 weeks назад @ paperswithcode.com
/joaojcorreia/ A Fuzzy Approach to Project Success: Measuring What Matters
/joaojcorreia/ A Fuzzy Approach to Project Success: Measuring What Matters

This paper introduces a novel approach to project success evaluation by integrating fuzzy logic into an existing construct. Code: https://github.com/joaojcorreia/FuzzyLogic_ProjectSuccess

5 months, 4 weeks назад @ paperswithcode.com
/kunkunlin1221/ InstructFLIP: Exploring Unified Vision-Language Model for Face Anti-spoofing
/kunkunlin1221/ InstructFLIP: Exploring Unified Vision-Language Model for Face Anti-spoofing

Extensive experiments demonstrate the effectiveness of InstructFLIP by outperforming SOTA models in accuracy and substantially reducing training redundancy across diverse domains in FAS. Code: https://github.com/kunkunlin1221/InstructFLIP

5 months, 4 weeks назад @ paperswithcode.com
/Linvyl/ Describe Anything Model for Visual Question Answering on Text-rich Images
/Linvyl/ Describe Anything Model for Visual Question Answering on Text-rich Images

Recent progress has been made in region-aware vision-language modeling, particularly with the emergence of the Describe Anything Model (DAM). Code: https://github.com/Linvyl/DAM-QA

5 months, 4 weeks назад @ paperswithcode.com
/abhijeet3922/ Developing Visual Augmented Q&A System using Scalable Vision Embedding Retrieval & Late Interaction Re-ranker
/abhijeet3922/ Developing Visual Augmented Q&A System using Scalable Vision Embedding Retrieval & Late Interaction Re-ranker

We propose multi-step custom implementation utilizing widely adopted hybrid search (metadata & embedding) and state of the art late interaction re-ranker to retrieve best matching pages. Code: https://github.com/abhijeet3922/vision-RAG

5 months, 4 weeks назад @ paperswithcode.com
/ziangcao0312/ PhysX: Physical-Grounded 3D Asset Generation
/ziangcao0312/ PhysX: Physical-Grounded 3D Asset Generation

3D modeling is moving from virtual to physical. Code: https://github.com/ziangcao0312/PhysX

5 months, 4 weeks назад @ paperswithcode.com
/henry123-boy/ SpatialTrackerV2: 3D Point Tracking Made Easy
/henry123-boy/ SpatialTrackerV2: 3D Point Tracking Made Easy

We present SpatialTrackerV2, a feed-forward 3D point tracking method for monocular videos. Code: https://github.com/henry123-boy/SpaTrackerV2

5 months, 4 weeks назад @ paperswithcode.com
/cncs-fit/ Emergence of Functionally Differentiated Structures via Mutual Information Optimization in Recurrent Neural Networks
/cncs-fit/ Emergence of Functionally Differentiated Structures via Mutual Information Optimization in Recurrent Neural Networks

Analysis of network performance, correlation patterns, and weight matrices reveals that mutual information minimization yields high task performance alongside clear functional modularity and moderate structural modularity. Code: https://github.com/cncs-fit/mio_rnn

5 months, 4 weeks назад @ paperswithcode.com
/coswindywang/ Making Language Model a Hierarchical Classifier and Generator
/coswindywang/ Making Language Model a Hierarchical Classifier and Generator

Language heads of the last layer are copied to different selected intermediate layers, and fine-tuned with different task inputs. Code: https://github.com/coswindywang/HdLM

5 months, 4 weeks назад @ paperswithcode.com
/ahmedehabb/ From Roots to Rewards: Dynamic Tree Reasoning with RL
/ahmedehabb/ From Roots to Rewards: Dynamic Tree Reasoning with RL

Modern language models address complex questions through chain-of-thought (CoT) reasoning (Wei et al., 2023) and retrieval augmentation (Lewis et al., 2021), yet struggle with error propagation and knowledge integration. Code: https://github.com/ahmedehabb/From-Roots-to-Rewards-Dynamic-Tree-Reasoning-with-RL

5 months, 4 weeks назад @ paperswithcode.com
💼 University and corporation labs
DeepMind DeepMind
последний пост 1 day назад
Veo 3.1 Ingredients to Video: More consistency, creativity and control
Veo 3.1 Ingredients to Video: More consistency, creativity and control Veo 3.1 Ingredients to Video: More consistency, creativity and control

Today, Veo is getting more expressive, with improvements that help you create more fun, creative, high-quality videos based on ingredient images, built directly for the mobile format.

We’re excited to bring new creative possibilities for everyone from casual storytellers to professional filmmakers.

We’re releasing:Improvements to Veo 3.1 Ingredients to Video, our capability that lets you create videos based on reference images.

This update makes videos more expressive and creative, even with simple prompts Native vertical outputs for Ingredients to Video (portrait mode) to power mobile-first, short-form video creation State-of-the-art upscaling to 1080p and 4K resolution for high-fidelity p…

1 day назад @ blog.google
Google's year in review: 8 areas with research breakthroughs in 2025
Google's year in review: 8 areas with research breakthroughs in 2025 Google's year in review: 8 areas with research breakthroughs in 2025

Google 2025 recap: Research breakthroughs of the year

3 weeks, 1 day назад @ deepmind.google
Gemini 3 Flash: frontier intelligence built for speed
Gemini 3 Flash: frontier intelligence built for speed Gemini 3 Flash: frontier intelligence built for speed

Today, we're expanding the Gemini 3 model family with the release of Gemini 3 Flash, which offers frontier intelligence built for speed at a fraction of the cost.

Last month, we kicked off Gemini 3 with Gemini 3 Pro and Gemini 3 Deep Think mode, and the response has been incredible.

With Gemini 3, we introduced frontier performance across complex reasoning, multimodal and vision understanding and agentic and vibe coding tasks.

Gemini 3 Flash retains this foundation, combining Gemini 3's Pro-grade reasoning with Flash-level latency, efficiency and cost.

Starting today, Gemini 3 Flash is rolling out to millions of people globally:

4 weeks назад @ blog.google
Gemma Scope 2: helping the AI safety community deepen understanding of complex language model behavior
Gemma Scope 2: helping the AI safety community deepen understanding of complex language model behavior Gemma Scope 2: helping the AI safety community deepen understanding of complex language model behavior

Today, we are releasing Gemma Scope 2: a comprehensive, open suite of interpretability tools for all Gemma 3 model sizes, from 270M to 27B parameters.

Producing Gemma Scope 2 involved storing approximately 110 Petabytes of data, as well as training over 1 trillion total parameters.

What’s new in Gemma Scope 2Interpretability research aims to understand the internal workings and learned algorithms of AI models.

Like its predecessor, Gemma Scope 2 acts as a microscope for the Gemma family of language models.

While the original Gemma Scope enabled research in key areas of safety, such as model hallucination, identifying secrets known by a model, and training safer models, Gemma Scope 2 support…

4 weeks, 1 day назад @ deepmind.google
Improved Gemini audio models for powerful voice experiences
Improved Gemini audio models for powerful voice experiences Improved Gemini audio models for powerful voice experiences

Earlier this week, we introduced greater control over audio generation with an upgrade to our Gemini 2.5 Pro and Flash Text-to-Speech models.

Today, we’re releasing an updated Gemini 2.5 Flash Native Audio for live voice agents.

Gemini 2.5 Flash Native Audio is now available across Google products including Google AI Studio, Vertex AI, and has also started rolling out in Gemini Live and Search Live, bringing the naturalness of native audio to Search Live for the first time.

Beyond powering helpful agents, native audio unlocks new possibilities for global communication.

Live Voice Agents

1 month назад @ blog.google
Deepening our partnership with the UK AI Security Institute
Deepening our partnership with the UK AI Security Institute Deepening our partnership with the UK AI Security Institute

Today, we're announcing an expanded partnership with the UK AI Security Institute (AISI) through a new Memorandum of Understanding focused on foundational security and safety research, to help ensure artificial intelligence is developed safely and benefits everyone.

The research partnership with AISI is an important part of our broader collaboration with the UK government on accelerating safe and beneficial AI progress.

This is why we have partnered with the UK AISI since its inception in November 2023 to test our most capable models.

We are actively working with AISI to build more robust evaluations for AI models, and our teams have collaborated on safety research to move the field forward…

1 month назад @ deepmind.google
Strengthening our partnership with the UK government to support prosperity and security in the AI era
Strengthening our partnership with the UK government to support prosperity and security in the AI era Strengthening our partnership with the UK government to support prosperity and security in the AI era

The UK has already laid a strong foundation to seize this moment and is uniquely positioned to translate AI innovation into public benefit.

That’s why we are excited to deepen our collaboration with the UK government to accelerate this work and offer a blueprint for other countries.

Accelerating access to frontier AI in key sectors: Science & EducationOur partnership will center on providing access to frontier AI in two areas foundational to the UK’s long-term success: scientific discovery and education.

The UK has a rich history of applying new technologies to drive scientific progress, from Hooke’s microscope to Faraday’s electrical experiments.

Establishing Google DeepMind’s first automa…

1 month назад @ deepmind.google
FACTS Benchmark Suite: Systematically evaluating the factuality of large language models
FACTS Benchmark Suite: Systematically evaluating the factuality of large language models FACTS Benchmark Suite: Systematically evaluating the factuality of large language models

The FACTS Benchmark SuiteToday, we’re teaming up with Kaggle to introduce the FACTS Benchmark Suite.

A Search Benchmark that tests a model’s ability to use Search as a tool to retrieve information and synthesize it correctly.

Similar to our previous release, we are following standard industry practice and keeping an evaluation set held-out as a private set.

The FACTS Benchmark Suite Score (or FACTS Score) is calculated as the average accuracy of both public and private sets across the four benchmarks.

Kaggle will oversee the management of the FACTS Benchmark Suite.

1 month назад @ deepmind.google
Engineering more resilient crops for a warming climate
Engineering more resilient crops for a warming climate Engineering more resilient crops for a warming climate

Scientists are using AlphaFold in their research to strengthen an enzyme that’s vital to photosynthesis, paving the way for more heat-tolerant crops.

As global warming accompanies more droughts and heatwaves, harvests of some staple crops are shrinking.

But less visible is what is happening inside these plants, where high heat can break down the molecular machinery that keeps them alive.

Plants use photosynthesis to produce the glucose that fuels their growth via an intricate choreography of enzymes inside plant cells.

"Our job is to learn from those examples and build that same resilience into the crops we depend on."

1 month, 1 week назад @ deepmind.google
AlphaFold: Five years of impact
AlphaFold: Five years of impact AlphaFold: Five years of impact

They used AlphaFold alongside comparative genomics to better understand how plants perceive changes in their environment, paving the way for more resilient crops.

AlphaFold has been cited in more than 35,000 papers and more than 200,000 papers incorporated elements of AlphaFold 2 in their methodology.

An independent analysis of AlphaFold 2’s impact, carried out by the Innovation Growth Lab, suggests that researchers using AlphaFold 2 see an increase of over 40% in their submission of novel experimental protein structures.

Those protein structures are more likely to be dissimilar to known structures, encouraging the exploration of uncharted areas of science.

The AlphaFold Server is empowerin…

1 month, 2 weeks назад @ deepmind.google
Revealing a key protein behind heart disease
Revealing a key protein behind heart disease Revealing a key protein behind heart disease

Both have a family history of heart disease – a reminder of what’s at stake in their work to better understand and ultimately help treat this deadly condition.

That protein, apoB100, has defied mapping not only because it’s enormous (for a protein), but also because it connects to fats and other molecules in complicated ways.

ApoB100 forms the molecular scaffold of “bad cholesterol”, which is known to scientists as low-density lipoprotein (LDL).

Discovering the structure of its key protein promised to shed light on how bad cholesterol becomes harmful inside the body, giving scientists a better chance to develop ways to prevent and treat ASCVD.

The images weren’t sharp enough to map the stru…

1 month, 2 weeks назад @ deepmind.google
Google DeepMind supports U.S. Department of Energy on Genesis: a national mission to accelerate innovation and scientific discovery
Google DeepMind supports U.S. Department of Energy on Genesis: a national mission to accelerate innovation and scientific discovery Google DeepMind supports U.S. Department of Energy on Genesis: a national mission to accelerate innovation and scientific discovery

We stand at an inflection point where the convergence of advanced AI and scientific research promises to unlock a new golden age of discovery.

There is perhaps no clearer expression of this than the application of AI within science.

Putting our advanced AI tools into the hands of American scientistsGoogle DeepMind will provide an accelerated access program for scientists at all 17 DOE National Laboratories to our frontier AI for Science models and agentic tools, starting today with AI co-scientist on Google Cloud.

We’re excited to see what America’s leading researchers will be able to do with our frontier AI models and agentic tools.

By combining human ingenuity with advanced AI capabilitie…

1 month, 3 weeks назад @ deepmind.google
How we’re bringing AI image verification to the Gemini app
How we’re bringing AI image verification to the Gemini app How we’re bringing AI image verification to the Gemini app

At Google, we’ve long invested in ways to provide you with helpful context about information you see online.

Now, as generative media becomes increasingly prevalent and high-fidelity, we are deploying tools to help you more easily determine whether the content you're interacting with was created or edited using AI.

Starting today, we’re making it easier for everyone to verify if an image was generated with or edited by Google AI right in the Gemini app, using SynthID, our digital watermarking technology that embeds imperceptible signals into AI-generated content.

Since then, over 20 billion AI-generated pieces of content have been watermarked using SynthID, and we have been testing our Synt…

1 month, 3 weeks назад @ blog.google
Build with Nano Banana Pro, our Gemini 3 Pro Image model
Build with Nano Banana Pro, our Gemini 3 Pro Image model Build with Nano Banana Pro, our Gemini 3 Pro Image model

Today, we’re releasing Nano Banana Pro (Gemini 3 Pro Image), a higher-fidelity model built on Gemini 3 Pro for developers to access studio-quality image generation.

This follows our release of Nano Banana (Gemini 2.5 Flash Image) just a few months ago.

Since then, we’ve loved seeing the community put its key features to work — from character consistency to photo restoration, and even using its capabilities to make local edits in an infinite canvas.

This state-of-the-art image generation and editing model is starting to roll out in paid preview to build a new wave of intelligent, multimodal applications with the Gemini API in Google AI Studio and Vertex AI for enterprises.

This model unlocks…

1 month, 3 weeks назад @ blog.google
Introducing Nano Banana Pro
Introducing Nano Banana Pro Introducing Nano Banana Pro

How Nano Banana Pro helps you bring any idea or design to lifeNano Banana Pro can help you visualize any idea and design anything — from prototypes, to representing data as infographics, to turning handwritten notes into diagrams.

With Nano Banana Pro, now you can:Generate more accurate, context-rich visuals based on enhanced reasoning, world knowledge and real-time informationWith Gemini 3’s advanced reasoning, Nano Banana Pro doesn’t just create beautiful images, it also helps you create more helpful content.

You can get accurate educational explainers to learn more about a new subject, like context-rich infographics and diagrams based on the content you provide or facts from the real wor…

1 month, 3 weeks назад @ blog.google
Google
последний пост 1 day назад
A gRPC transport for the Model Context Protocol
A gRPC transport for the Model Context Protocol A gRPC transport for the Model Context Protocol

An interesting alternative to MCP transcoding is to use gRPC as the native transport for MCP.

Specifically, we’re committed to supporting gRPC practitioners in their journey to adopt MCP in production, and we’re actively working with the MCP community to explore mechanisms to support gRPC as a transport for MCP.

A community-backed transport package will enable gRPC practitioners to deploy MCP with gRPC in a consistent and interoperable manner.

The native use of gRPC as a transport avoids the need for transcoding and helps maintain operational consistency for environments that are actively using gRPC.

In the rest of this post, we explore the benefits of using gRPC as a native transport for M…

1 day назад @ cloud.google.com
Build data analytics agents faster with BigQuery’s fully managed, remote MCP server
Build data analytics agents faster with BigQuery’s fully managed, remote MCP server Build data analytics agents faster with BigQuery’s fully managed, remote MCP server

With the release of fully managed, remote Model Context Protocol (MCP) servers for Google services last month, you can now use BigQuery MCP server to give your AI agents a direct, secure, way to analyze data.

This fully managed MCP server removes management overhead, enabling you to focus on developing intelligent agents.

In this blog post, we discuss and demonstrate the integrations of newly released fully managed, remote BigQuery Server, which is in preview as of January 2026.

Remote MCP servers run on the service's infrastructure and offer an HTTP endpoint to AI applications.

This enables communication between the AI MCP client and the MCP server using a defined standard.

1 week назад @ cloud.google.com
Cloud CISO Perspectives: 2025 in review: Cloud security basics and evolving AI
Cloud CISO Perspectives: 2025 in review: Cloud security basics and evolving AI Cloud CISO Perspectives: 2025 in review: Cloud security basics and evolving AI

Building the most trusted cloudWe continued to enhance our security capabilities and controls on our cloud platform to help organizations secure their cloud environments and address evolving policy, compliance, and business objectives.

Our forecast for 2026As security professionals, we know that threat actors will continue to innovate to achieve their mission objectives.

To help defenders proactively prepare for the coming year, we publish our annual forecast report with insights from across Google.

We look forward to sharing more insights to help organizations strengthen their security posture in the new year.

For more leadership guidance from Google Cloud experts, please visit our CISO In…

3 weeks, 5 days назад @ cloud.google.com
Getting AI to write good SQL: Optimizing the AlloyDB AI natural language API for your use case
Getting AI to write good SQL: Optimizing the AlloyDB AI natural language API for your use case Getting AI to write good SQL: Optimizing the AlloyDB AI natural language API for your use case

Descriptive and prescriptive contextAs mentioned above, the AlloyDB AI natural language API relies on descriptive and prescriptive context to improve the accuracy of the SQL code it generates.

The AlloyDB AI natural language API facilitates the creation of descriptive and prescriptive context.

The value index clarifies what kind of entity “John Smith” is, and can be automatically created by AlloyDB AI for your application.

Natural language search over structured, unstructured and multimodal dataWhen it comes to applications that provide search over structured data, the AlloyDB AI natural language API enables a clean and powerful search experience.

Bringing the AlloyDB AI natural language AP…

3 weeks, 5 days назад @ cloud.google.com
Announcing advanced governance capabilities for Vertex AI Agent Builder
Announcing advanced governance capabilities for Vertex AI Agent Builder Announcing advanced governance capabilities for Vertex AI Agent Builder

At Google Cloud, we continue to make critical investments to Vertex AI Agent Builder, our comprehensive and open platform, enabling you to build faster, scale efficiently, and govern with enterprise-grade security.

Today, with the integration of the Cloud API Registry, we’re excited to bring enhanced tool governance capabilities to Vertex AI Agent Builder.

With this latest update, administrators can now manage available tools for developers across your organization directly in Vertex AI Agent Builder Console, and developers can leverage tools managed by the registry with a new ApiRegistry .

Govern your tools with confidenceBuilding a useful agent requires the agent to have access to the nec…

3 weeks, 6 days назад @ cloud.google.com
Automate AI and HPC clusters with Cluster Director, now generally available
Automate AI and HPC clusters with Cluster Director, now generally available Automate AI and HPC clusters with Cluster Director, now generally available

Today, we are delivering on those requirements with the General Availability (GA) of Cluster Director and the Preview of Cluster Director support for Slurm on Google Kubernetes Engine (GKE).

Cluster Director (GA) is a managed infrastructure service designed to meet the rigorous demands of modern supercomputing.

There's no extra charge to use Cluster Director.

How Cluster Director supports each phase of deploymentDay 0: PreparationStanding up a cluster typically involves weeks of planning, wrangling Terraform, and debugging the network.

Cluster Director changes the ‘Day 0’ experience entirely, with tools for designing infrastructure topology that’s optimized for your workload requirements.

3 weeks, 6 days назад @ cloud.google.com
Google named a Leader in The Forrester Wave™: AI Infrastructure Solutions, Q4 2025
Google named a Leader in The Forrester Wave™: AI Infrastructure Solutions, Q4 2025 Google named a Leader in The Forrester Wave™: AI Infrastructure Solutions, Q4 2025

Yesterday, Forrester released The Forrester Wave™: AI Infrastructure Solutions, Q4 2025 report, evaluating 13 vendors, and we believe their findings validate our commitment to solving these core challenges.

Access the full report: The Forrester Wave™: AI Infrastructure Solutions, Q4 2025Accelerating time-to-value with an integrated systemEnterprises don’t run AI in a vacuum.

Delivering continuous AI innovationWe are honored to be recognized as a Leader in The Forrester Wave™ report, which we believe validates decades of R&D and our approach to building ultra-scale AI infrastructure.

Access the full report: The Forrester Wave™: AI Infrastructure Solutions, Q4 20251.

IDC Business Value Snapsh…

4 weeks назад @ cloud.google.com
Introducing Gemini 3 Flash: Intelligence and speed for enterprises
Introducing Gemini 3 Flash: Intelligence and speed for enterprises Introducing Gemini 3 Flash: Intelligence and speed for enterprises

Today, we’re expanding the Gemini 3 model family with Gemini 3 Flash, which offers frontier intelligence built for speed at a fraction of the cost.

Gemini 3 Flash builds on the model series that developers and enterprises already love, optimized for high-frequency workflows that demand speed, without sacrificing quality.

Gemini 3 Flash is built to be highly efficient, pushing the boundaries of quality at better price performance and faster speed.

It is available now in Gemini Enterprise, Vertex AI, and Gemini CLI, so businesses and developers can access:Advanced multimodal processing: Gemini 3 Flash enables enterprises to build applications capable of complex video analysis, data extraction…

4 weeks назад @ cloud.google.com
Connect your enterprise data to Google’s new Antigravity IDE
Connect your enterprise data to Google’s new Antigravity IDE Connect your enterprise data to Google’s new Antigravity IDE

Google Cloud is at the forefront of this shift, empowering you to build robust, data-driven applications quickly and accurately.

With Model Context Protocol (MCP) servers powered by MCP Toolbox for Databases now available within Antigravity, you can securely connect your AI agents to services like AlloyDB for PostgreSQL, BigQuery, Spanner, Cloud SQL, Looker and others within Google’s Data Cloud, all within your development workflow.

We designed Antigravity to keep you in the flow, but the power of an AI agent is limited by what it "knows."

By integrating pre-built MCP servers directly into Antigravity, you don’t need to perform any manual configuration.

Discover and launchYou can find MCP s…

1 month назад @ cloud.google.com
A developer's guide to Gemini Live API in Vertex AI
A developer's guide to Gemini Live API in Vertex AI A developer's guide to Gemini Live API in Vertex AI

Today, we announced the general availability of Gemini Live API on Vertex AI, which is powered by the latest Gemini 2.5 Flash Native Audio model.

In this post we'll look at two templates and three reference demos that help you understand how to best use Gemini Live API.

Gemini Live API fundamentally changes the engineering approach with a unified, low-latency, native audio architecture.

Native audio processing: Gemini 2.5 Flash Native Audio model processes raw audio natively through a single, low-latency model.

Next-generation conversation featuresGemini Live API gives you a suite of production-ready features that define a new standard for AI agents:

1 month назад @ cloud.google.com
Cloud CISO Perspectives: Our 2026 Cybersecurity Forecast report
Cloud CISO Perspectives: Our 2026 Cybersecurity Forecast report Cloud CISO Perspectives: Our 2026 Cybersecurity Forecast report

Marina Kaganovich, executive trust leadThe heightened capability of agentic AI to take actions and execute tasks autonomously elevates the importance of cybersecurity basics.

Vesselin Tzvetkov, senior cybersecurity advisorAs Francis noted, agentic security operations are set to become the standard for modern SOCs, dramatically enhancing the speed and capabilities of security organizations.

Vinod D’Souza, director, manufacturing and industryIn 2026, agentic AI will help the manufacturing and industrial sector cross the critical threshold from static automation to true autonomy.

By rooting security strategies in data-centered Zero Trust, organizations stop treating security as a gatekeeper an…

1 month назад @ cloud.google.com
How to connect Looker to Gemini Enterprise in minutes, with MCP Toolbox and ADK
How to connect Looker to Gemini Enterprise in minutes, with MCP Toolbox and ADK How to connect Looker to Gemini Enterprise in minutes, with MCP Toolbox and ADK

We can all agree that the quality of AI-driven answers relies on the consistency of the underlying data.

Building off the recent introduction of Looker’s Model Context Protocol (MCP) server, in this blog we take you through the process of creating an Agent Development Kit (ADK) agent that is connected to Looker via the MCP Toolbox for Databases and exposing it within Gemini Enterprise.

Instead of managing tool logic and authentication themselves, agents act as MCP clients and request tools from the Toolbox.

The MCP Toolbox handles all the underlying complexities, including secure connections to Looker, authentication and query execution.

The MCP Toolbox for Databases natively supports Looke…

1 month назад @ cloud.google.com
Gemini Live API Now GA on Vertex AI
Gemini Live API Now GA on Vertex AI Gemini Live API Now GA on Vertex AI

Today, we are excited to announce that Gemini Live API, powered by the latest Gemini 2.5 Flash Native Audio model, is generally available on Vertex AI.

A new standard with real-time multimodal AI agentsGemini Live API represents a new standard for bringing AI to life.

Deploying on Vertex AI allows you to leverage our expanding global infrastructure across multiple regions, delivering reliability for your users.

Building real-world impact with Gemini Live APIThe true power of Gemini Live API is demonstrated by the companies who are using it today to redefine their customer experiences.

Shopify, the leading global commerce platform, developed Sidekick, a multimodal AI assistant powered by Gem…

1 month назад @ cloud.google.com
AI agents are here. Is your infrastructure ready?
AI agents are here. Is your infrastructure ready? AI agents are here. Is your infrastructure ready?

In a recent IDC global survey of over 1,300 AI decision-makers, inference was already cited as the largest AI workload segment, accounting for 47% of all AI operations.

This surge in demand is exposing a critical vulnerability for many organizations: the AI efficiency gap.

The TCO crisis in an age of agentsThe AI efficiency gap is the difference between the theoretical performance of an AI stack and the actual, real-world performance achieved.

That is why we created AI Hypercomputer: an integrated supercomputer system designed to deliver exceptional performance and efficiency for demanding AI workloads.

Get your free copy of the whitepaper to learn more: The AI Efficiency Gap: From TCO Cris…

1 month назад @ cloud.google.com
How we built a multi-agent system for superior business forecasting
How we built a multi-agent system for superior business forecasting How we built a multi-agent system for superior business forecasting

This innovative solution combines two powerful, specialized AI agents: a prediction agent built by Google Cloud and App Orchid’s Data Agent offering.

Google prediction agent - The forecasting powerhouseThe prediction agent, which is primarily the custom engineering work of Google Cloud, is the system’s window to the future.

App Orchid Data Agent - The enterprise intelligence data expertAccurate predictions depend on high-quality, AI-ready data, which is where App Orchid’s Data Agent excels.

The combined business forecasting agentAt the heart of the solution is a unified business forecasting agent, which brings together the capabilities of our unique prediction and data agents in a discrete …

1 month назад @ cloud.google.com
OpenAI
последний пост None
Microsoft Microsoft
последний пост 1 month назад
Agent Lightning: Adding reinforcement learning to AI agents without code rewrites
Agent Lightning: Adding reinforcement learning to AI agents without code rewrites Agent Lightning: Adding reinforcement learning to AI agents without code rewrites

To address this, a research team from Microsoft Research Asia – Shanghai has introduced Agent Lightning.

Whether it involves multiple collaborating agents or dynamic tool use, Agent Lightning breaks it down into a sequence of transitions.

Agent Lightning as middlewareAgent Lightning serves as middleware between RL algorithms and agent environments, providing with modular components that enable scalable RL through standardized protocols and well-defined interfaces.

In practice, developers can keep their existing agent frameworks and switch model calls to the Agent Lightning API without changing their agent code (Figure 5).

By bridging existing agentic systems with reinforcement learning, Age…

1 month назад @ microsoft.com
Promptions helps make AI prompting more precise with dynamic UI controls
Promptions helps make AI prompting more precise with dynamic UI controls Promptions helps make AI prompting more precise with dynamic UI controls

To address this, we are excited to introduce Promptions (prompt + options), a UI framework that helps developers build AI interfaces with more precise user control.

We compared the static design from the first study, called the “Static Prompt Refinement Control” (Static PRC), against a “Dynamic Prompt Refinement Control” (Dynamic PRC) with features that responded to participants’ feedback.

Comparison of user preferences for Static PRC versus Dynamic PRC across key evaluation criteria.

(1) The Option Module reads the user’s prompt and conversation history and (2) generates prompt options.

Key usability challenges include clarifying how dynamic options affect AI output and managing the comple…

1 month назад @ microsoft.com
GigaTIME: Scaling tumor microenvironment modeling using virtual population generated by multimodal AI
GigaTIME: Scaling tumor microenvironment modeling using virtual population generated by multimodal AI GigaTIME: Scaling tumor microenvironment modeling using virtual population generated by multimodal AI

C, Scatter plot comparing the subtype-level GigaTIME-translated virtual mIF activations between TCGA and Providence virtual populations.

To our knowledge, this is the first large-scale study exploring multimodal AI for scaling virtual mIF generation.

H, A case study showcasing the activation maps across different virtual mIF channels for a H&E slide in our virtual population, and virtual mIF of sample patches from this slide.

By applying GigaTIME to Providence real-world data, we generated a virtual population of 14,256 patients with virtual mIF and key clinical attributes.

G, Bar plot comparing pan-cancer patient stratification performance in terms of survival log rank p-values among virtu…

1 month назад @ microsoft.com
Ideas: Community building, machine learning, and the future of AI
Ideas: Community building, machine learning, and the future of AI Ideas: Community building, machine learning, and the future of AI

This week, machine learning researchers around the world will be attending the annual Conference on Neural Information Processing Systems, or NeurIPS.

In this series, we’ll explore the technologies that are shaping our future and the big ideas that propel them forward.

So around that time when I started my PhD at Penn, I was working in machine learning theory and algorithmic economics.

How had you experienced a lack of community or network of women in machine learning before the founding of WiML?

So particularly when working on topics related to fairness, I’ve ended up focusing a bunch on stuff to do with marginalized groups as part of my responsible AI work.

1 month, 1 week назад @ microsoft.com
Reducing Privacy leaks in AI: Two approaches to contextual integrity
Reducing Privacy leaks in AI: Two approaches to contextual integrity Reducing Privacy leaks in AI: Two approaches to contextual integrity

The theory of contextual integrity frames privacy as the appropriateness of information flow within specific social contexts.

Each tackles contextual integrity from a different angle, but both aim to build directly into AI systems a greater sensitivity to information-sharing norms.

Contextual Integrity in LLMs via Reasoning and Reinforcement Learning, accepted at NeurIPS 2025, takes a different approach to applying contextual integrity.

Contextual integrity through reasoning and reinforcement learningIn our second paper, we explore whether contextual integrity can be built into the model itself rather than enforced through external checks at inference time.

To address this trade-off, we int…

1 month, 2 weeks назад @ microsoft.com
Fara-7B: An Efficient Agentic Model for Computer Use
Fara-7B: An Efficient Agentic Model for Computer Use Fara-7B: An Efficient Agentic Model for Computer Use

Today, we are pleased to announce Fara-7B, our first agentic SLM designed specifically for computer use.

Unlike traditional chat models that generate text-based responses, Computer Use Agent (CUA) models like Fara-7B leverage computer interfaces, such as a mouse and keyboard, to complete tasks on behalf of users.

This results in reduced latency and improved privacy, as user data remains local.

Fara-7B breaks ground on a new pareto frontier, showing that on-device computer use agents are approaching the capabilities of frontier models.

For guidance on how to use our model safely, and the security considerations to be mindful of when using our model, please refer to our Model card (opens in n…

1 month, 2 weeks назад @ microsoft.com
MMCTAgent: Enabling multimodal reasoning over large video and image collections
MMCTAgent: Enabling multimodal reasoning over large video and image collections MMCTAgent: Enabling multimodal reasoning over large video and image collections

Real-world reasoning increasingly involves analyzing long-form video content, where context spans minutes or hours, far beyond the context limits of most models.

The Planner agent decomposes a user query, identifies the appropriate reasoning tools, performs multimodal operations, and drafts a preliminary answer.

MMCTAgent’s Planner–Critic architecture enables multimodal reasoning over long-form video through structured ingestion, retrieval, and iterative feedback.

The VideoAgent extends this architecture to long-form video reasoning.

Takeaways and next stepsMMCTAgent demonstrates a scalable agentic approach to multimodal reasoning with a Planner–Critic architecture.

2 months назад @ microsoft.com
BlueCodeAgent: A blue teaming agent enabled by automated red teaming for CodeGen AI
BlueCodeAgent: A blue teaming agent enabled by automated red teaming for CodeGen AI BlueCodeAgent: A blue teaming agent enabled by automated red teaming for CodeGen AI

Many studies have explored red teaming code LLMs, testing whether the models can reject unsafe requests and whether their generated code exhibits insecure patterns.

Knowledge-enhanced blue teaming: Building on the foundation of red-teaming knowledge, BlueCodeAgent significantly improves blue-teaming performance by leveraging constitutions derived from knowledge and dynamic testing.

Generalization to seen and unseen risks: Empowered by comprehensive red-teaming knowledge, BlueCodeAgent generalizes effectively to unseen risks.

A blue teaming agent enabled by red teamingFigure 2: Overview of BlueCodeAgent, an end-to-end blue teaming framework powered by automated red teaming for code security.…

2 months назад @ microsoft.com
When industry knowledge meets PIKE-RAG: The innovation behind Signify’s customer service boost
When industry knowledge meets PIKE-RAG: The innovation behind Signify’s customer service boost When industry knowledge meets PIKE-RAG: The innovation behind Signify’s customer service boost

Spotlight: Event Series Microsoft Research Forum Join us for a continuous exchange of ideas about research in the era of general AI.

These differentiated advantages stem from PIKE-RAG’s unique approach to understanding and processing professional knowledge.

“It’s also worth noting that the researchers at Microsoft Research Asia demonstrated strong industry knowledge and rigorous scientific methodology.

Through this collaboration, we validated that PIKE-RAG’s general approach can greatly improve the accuracy of professional knowledge Q&A and accelerate scenario customization.

Our researchers also gained valuable experience in handling domain-specific data,” explained Jiang Bian, partner rese…

2 months, 1 week назад @ microsoft.com
Magentic Marketplace: an open-source simulation environment for studying agentic markets
Magentic Marketplace: an open-source simulation environment for studying agentic markets Magentic Marketplace: an open-source simulation environment for studying agentic markets

To help navigate this uncertainty, we built Magentic Marketplace (opens in new tab)— an open-source simulation environment for exploring the numerous possibilities of agentic markets and their societal implications at scale.

To explore these dynamics in depth, the Magentic Marketplace platform enables controlled experimentation across diverse agentic marketplace scenarios.

With Magentic Marketplace, researchers can model how agents representing customers and businesses interact—shedding light on the dynamics that could shape future digital markets.

Magentic Marketplace includes two agent types: Assistant Agents (customers) and Service Agents (businesses).

Unlike traditional markets, which d…

2 months, 1 week назад @ microsoft.com
RedCodeAgent: Automatic red-teaming agent against diverse code agents
RedCodeAgent: Automatic red-teaming agent against diverse code agents RedCodeAgent: Automatic red-teaming agent against diverse code agents

In the context of code, effective red-teaming requires more than simply checking whether the target code agent rejects unsafe requests.

After the second request was rejected by the code agent, RedCodeAgent invoked both Code Substitution and GCG to optimize the prompt.

Ultimately, RedCodeAgent successfully combined the suggestion from Code Substitution (i.e., using pathlib) with the adversarial suffix generated by GCG, making the target code agent delete the specified file.

In the context of code, it is not enough for the target code agent to simply avoid rejecting the request; the target code agent must also generate and execute code that performs the intended function.

Quantitatively, we f…

2 months, 1 week назад @ microsoft.com
Tell me when: Building agents that can wait, monitor, and act
Tell me when: Building agents that can wait, monitor, and act Tell me when: Building agents that can wait, monitor, and act

This matters because monitoring tasks are everywhere.

To address this, we are introducing SentinelStep (opens in new tab), a mechanism that enables agents to complete long-running monitoring tasks.

Most real-world monitoring tasks share this limitation, making systematic bench marking very challenging.

In response, we are developing SentinelBench, a suite of synthetic web environments for evaluating monitoring tasks.

By embedding patience into plans, agents can responsibly monitor conditions and act when it matters—staying proactive without wasting resources.

2 months, 3 weeks назад @ microsoft.com
Ideas: More AI-resilient biosecurity with the Paraphrase Project
Ideas: More AI-resilient biosecurity with the Paraphrase Project Ideas: More AI-resilient biosecurity with the Paraphrase Project

Today, I’m excited to talk about the Paraphrase Project, an effort I co-led exploring how advances in AI tools for protein design might impact biosecurity.

These “patches,” akin to those in cybersecurity, have now been shared with organizations globally to strengthen biosecurity screening.

The project highlights that the same AI tools capable of incredible good can also be misused, requiring us to be vigilant, thoughtful, and creative so we continue to get the most benefit out of AI tools while working to ensure that we avoid costly misuses.

So things like, how similar is this to that template, wild-type protein structure that we used as our conditioning information?

But I feel like broadly…

3 months, 1 week назад @ microsoft.com
Ideas: More AI-resilient biosecurity with the Paraphrase Project
Ideas: More AI-resilient biosecurity with the Paraphrase Project Ideas: More AI-resilient biosecurity with the Paraphrase Project

Today, I’m excited to talk about the Paraphrase Project, an effort I co-led exploring how advances in AI tools for protein design might impact biosecurity.

These “patches,” akin to those in cybersecurity, have now been shared with organizations globally to strengthen biosecurity screening.

The project highlights that the same AI tools capable of incredible good can also be misused, requiring us to be vigilant, thoughtful, and creative so we continue to get the most benefit out of AI tools while working to ensure that we avoid costly misuses.

So things like, how similar is this to that template, wild-type protein structure that we used as our conditioning information?

But I feel like broadly…

3 months, 1 week назад @ microsoft.com
When AI Meets Biology: Promise, Risk, and Responsibility
When AI Meets Biology: Promise, Risk, and Responsibility When AI Meets Biology: Promise, Risk, and Responsibility

In computer-based studies, we found that AI protein design (AIPD) tools could generate modified versions of proteins of concern, such as ricin.

Azure AI Foundry Labs Get a glimpse of potential future directions for AI, with these experimental technologies from Microsoft Research.

Stratified tiers of information : Data and code are classified into several tiers according to their potential hazard, from low-risk summaries through sensitive technical data to critical software pipelines.

The Age of AI in the Life Sciences: Benefits and Biosecurity Considerations, National Academies of Science, Engineering, and Medicine, 2025.

(opens in new tab)Protecting scientific integrity in an age of genera…

3 months, 1 week назад @ microsoft.com
MIT AI MIT AI
последний пост 5 days, 12 hours назад
3 Questions: How AI could optimize the power grid
3 Questions: How AI could optimize the power grid 3 Questions: How AI could optimize the power grid

Q: Why does the power grid need to be optimized in the first place?

Q: How can AI be most useful in power grid optimization?

This could lead to a cleaner power grid by allowing us to handle and better utilize these resources.

I’m excited to develop AI algorithms that respect the physical constraints of the power grid so that we can credibly deploy them.

But if you make the same magnitude of a mistake when you are optimizing a power grid, that can cause a large-scale blackout.

5 days, 12 hours назад @ news.mit.edu
Decoding the Arctic to predict winter weather
Decoding the Arctic to predict winter weather Decoding the Arctic to predict winter weather

Cohen, a research scientist in MIT’s Department of Civil and Environmental Engineering (CEE), has spent decades studying how conditions in the Arctic set the course for winter weather throughout Europe, Asia, and North America.

AI subseasonal forecastingWhile AI weather models have made impressive strides showcasing in short-range (one-to–10-day) forecasts, these advances have not yet applied to longer periods.

That gap is why this year could be a turning point for subseasonal weather forecasting.

If validated, Cohen explains, it would show how combining Arctic indicators with AI could extend the lead time for predicting impactful weather.

In November, Cohen even appeared as a clue in The W…

5 days, 19 hours назад @ news.mit.edu
Stone Center on Inequality and Shaping the Future of Work Launches at MIT
Stone Center on Inequality and Shaping the Future of Work Launches at MIT Stone Center on Inequality and Shaping the Future of Work Launches at MIT

The James M. and Cathleen D. Stone Center on Inequality and Shaping the Future of Work officially launched on Nov. 3, 2025, bringing together scholars, policymakers, and practitioners to explore critical questions about economic opportunity, technology, and democracy.

Co-directed by MIT professors Daron Acemoglu, David Autor, and Simon Johnson, the new Stone Center analyzes the forces that contribute to growing income and wealth inequality through the erosion of job quality and labor market opportunities for workers without a college degree.

MIT Provost Anantha Chandrakasan opened the launch event by emphasizing the urgency and importance of the center's mission.

To mitigate wealth inequali…

6 days, 21 hours назад @ news.mit.edu
MIT scientists investigate memorization risk in the age of clinical AI
MIT scientists investigate memorization risk in the age of clinical AI MIT scientists investigate memorization risk in the age of clinical AI

What is patient privacy for?

Foundation models trained on EHRs should normally generalize knowledge to make better predictions, drawing upon many patient records.

But in “memorization,” the model draws upon a singular patient record to deliver its output, potentially violating patient privacy.

Notably, foundation models are already known to be prone to data leakage.

“There’s a reason our health data is private,” Tonekaboni says.

1 week, 1 day назад @ news.mit.edu
Using design to interpret the past and envision the future
Using design to interpret the past and envision the future Using design to interpret the past and envision the future

He is a collaborator at the Design Intelligence Lab and has served as a teaching assistant in MIT’s architecture wood shop, helping students to bring together digital design techniques with hands-on fabrication.

Following the class, Payne continued working on models and drawings reconstructing some important Tuskegee architecture.

Incorporating AI to design for the futureWhile much of Payne’s research is rooted in the past, he is also interested in artificial intelligence and its implications for future innovations.

In addition, Payne took a class about large language objects taught by associate professor of the practice Marcelo Coelho, director of the Design Intelligence Lab.

Payne now con…

1 week, 1 day назад @ news.mit.edu
MIT in the media: 2025 in review
MIT in the media: 2025 in review MIT in the media: 2025 in review

“At MIT, innovation ranges from awe-inspiring technology to down-to-Earth creativity,” noted Chronicle, during a campus visit this year for an episode of the program.

In 2025, MIT researchers made headlines across print publications, podcasts, and video platforms for key scientific advances, from breakthroughs in quantum and artificial intelligence to new efforts aimed at improving pediatric health care and cancer diagnosis.

Neha Narula, director of the MIT Digital Currency Initiative, examines the future of cash as the use of digital currencies expands.

Full story via Michigan Farm NewsBug-sized robots could help pollination on future farmsInsect-sized robots crafted by MIT researchers cou…

3 weeks, 1 day назад @ news.mit.edu
Guided learning lets “untrainable” neural networks realize their potential
Guided learning lets “untrainable” neural networks realize their potential Guided learning lets “untrainable” neural networks realize their potential

Remarkably, even untrained networks contain architectural biases that can be transferred, while trained guides additionally convey learned patterns.

This result underscores a key insight: Untrained networks already encode valuable architectural biases that can steer other networks toward effective learning.

By aligning with a guide network, it’s possible to separate the contributions of architectural biases from those of learned knowledge.

By revealing the hidden potential of even the most stubborn networks, guidance provides a powerful new tool for understanding — and hopefully shaping — the foundations of machine learning.

Remarkably, the authors show this can be done using small, untrain…

3 weeks, 5 days назад @ news.mit.edu
A new way to increase the capabilities of large language models
A new way to increase the capabilities of large language models A new way to increase the capabilities of large language models

Kim’s co-authors include lead author Songlin Yang, an EECS graduate student and former MIT-IBM Watson AI Lab Summer Program intern; Kaiyue Wen of Stanford University; Liliang Ren of Microsoft; and Yikang Shen, Shawn Tan, Mayank Mishra, and Rameswar Panda of IBM Research and the MIT-IBM Watson AI Lab.

The cumulative effect lets the system model how the meaning changes along the path between words, not just how far apart they are.

PaTH Attention improved perplexity and outcompeted other methods on reasoning benchmarks it wasn’t trained on.

PaTH Attention consistently proved capable of content-awareness.

In this way, PaTH Attention extends the expressive power of transformer architectures.

3 weeks, 6 days назад @ news.mit.edu
A “scientific sandbox” lets researchers explore the evolution of vision systems
A “scientific sandbox” lets researchers explore the evolution of vision systems A “scientific sandbox” lets researchers explore the evolution of vision systems

This framework could enable scientists to probe “what-if” questions about vision systems that are difficult to study experimentally.

Building a scientific sandboxThe paper began as a conversation among the researchers about discovering new vision systems that could be useful in different fields, like robotics.

Over many generations, agents evolve different elements of vision systems that maximize rewards.

Testing hypothesesWhen the researchers set up experiments in this framework, they found that tasks had a major influence on the vision systems the agents evolved.

In the future, the researchers want to use this simulator to explore the best vision systems for specific applications, which c…

3 weeks, 6 days назад @ news.mit.edu
“Robot, make me a chair”
“Robot, make me a chair” “Robot, make me a chair”

The researchers tackled these challenges using a vision-language model (VLM), a powerful generative AI model that has been pre-trained to understand images and text.

By serving as both the eyes and brain of the robot, the VLM enables the robot to do this,” Kyaw says.

A user prompts the system with text, perhaps by typing “make me a chair,” and gives it an AI-generated image of a chair to start.

For instance, the model can determine that the seat and backrest should have panels to have surfaces for someone sitting and leaning on the chair.

Then the VLM chooses the labels that correspond to the geometric parts of the chair that should receive panels on the 3D mesh to complete the design.

4 weeks, 1 day назад @ news.mit.edu
3 Questions: Using computation to study the world’s best single-celled chemists
3 Questions: Using computation to study the world’s best single-celled chemists 3 Questions: Using computation to study the world’s best single-celled chemists

Q: What drew you to research microbes in extreme environments, and what are the challenges in studying them?

I wanted to be an astronaut growing up, and the closest thing to astrobiology is examining extreme environments on Earth.

And the only thing that lives in those extreme environments are microbes.

My latest work is genomic language modeling.

A genomic language model is technically a large language model, except the language is DNA as opposed to human language.

4 weeks, 1 day назад @ news.mit.edu
Working to eliminate barriers to adopting nuclear energy
Working to eliminate barriers to adopting nuclear energy Working to eliminate barriers to adopting nuclear energy

What if there were a way to solve one of the most significant obstacles to the use of nuclear energy — the disposal of high-level nuclear waste (HLW)?

Such a move would be especially important for the public’s acceptance of nuclear energy.

“We’re reframing the problem of nuclear waste, transforming it from a liability to an energy source,” Sarsenbayev says.

The nuances of nuclearSarsenbayev had to do a bit of reframing himself in how he perceived nuclear energy.

Removing the bottleneck for nuclear energy adoption by producing carbon-free power and ensuring the safe disposal of radioactive waste.

4 weeks, 1 day назад @ news.mit.edu
Deep-learning model predicts how fruit flies form, cell by cell
Deep-learning model predicts how fruit flies form, cell by cell Deep-learning model predicts how fruit flies form, cell by cell

In a study appearing today in the journal Nature Methods, the team presents a new deep-learning model that learns, then predicts, how certain geometric properties of individual cells will change as a fruit fly develops.

The team applied the model to videos of developing fruit fly embryos, each of which starts as a cluster of about 5,000 cells.

As a proof of principle, the team trained the new model to “learn” how individual cells change over time during fruit fly gastrulation.

“The overall shape of the fruit fly at this stage is roughly an ellipsoid, but there are gigantic dynamics going on at the surface during gastrulation,” Guo says.

What’s more, the videos contain labels of individual c…

1 month назад @ news.mit.edu
Enabling small language models to solve complex reasoning tasks
Enabling small language models to solve complex reasoning tasks Enabling small language models to solve complex reasoning tasks

As language models (LMs) improve at tasks like image generation, trivia questions, and simple math, you might think that human-like reasoning is around the corner.

Small LMs can’t do this reliably on their own; large language models (LLMs) sometimes can, particularly if they’re optimized for reasoning tasks, but they take a while to respond, and they use a lot of computing power.

Then, the LLM relays these instructions and guidelines in a clear way to smaller models.

For instance, whereas existing reasoning models like OpenAI’s o1 perform reasoning in text, DisCIPL “reasons” by writing Python code, which is more compact.

DisCIPL’s efficiency gains stem partly from using small Llama models a…

1 month назад @ news.mit.edu
New MIT program to train military leaders for the AI age
New MIT program to train military leaders for the AI age New MIT program to train military leaders for the AI age

Artificial intelligence can enhance decision-making and enable action with reduced risk and greater precision, making it a critical tool for national security.

“The potential for artificial intelligence is just starting to be fully realized.

The 2N6 curriculum is application focused, and the content is built to satisfy the U.S. Navy’s sub-specialty code for Applied Artificial Intelligence.

“The admiral made the connection, envisioning an applied AI program similar to 2N.”2N6 will run as a pilot program for at least two years.

The program’s first cohort will comprise only U.S. Navy officers, with plans to expand more broadly.

1 month назад @ news.mit.edu
Berkeley AI
последний пост 2 months, 2 weeks назад
RL without TD learning
RL without TD learning RL without TD learning

RL without TD learningIn this post, I’ll introduce a reinforcement learning (RL) algorithm based on an “alternative” paradigm: divide and conquer.

We can do Reinforcement Learning (RL) based on divide and conquer, instead of temporal difference (TD) learning.

There are two classes of algorithms in RL: on-policy RL and off-policy RL.

We compared TRL with $n$-step TD learning with different values of $n$, from $1$ (pure TD) to $\infty$ (pure MC).

I still think one of the most important problems in RL (and even in machine learning) is to find a scalable off-policy RL algorithm.

2 months, 2 weeks назад @ bair.berkeley.edu
What exactly does word2vec learn?
What exactly does word2vec learn? What exactly does word2vec learn?

What exactly does word2vec learn?

What exactly does word2vec learn, and how?

In this framing, it’s clear that word2vec is a minimal neural language model.

As a result, the theory predicts exactly what features are learned in terms of the corpus statistics and the algorithmic hyperparameters.

We find that over the course of learning, word2vec builds these linear representations in a sequence of noisy learning steps, and their geometry is well-described by a spiked random matrix model.

4 months, 2 weeks назад @ bair.berkeley.edu
Whole-Body Conditioned Egocentric Video Prediction
Whole-Body Conditioned Egocentric Video Prediction Whole-Body Conditioned Egocentric Video Prediction

Whole-Body Conditioned Egocentric Video Prediction×Predicting Ego-centric Video from human Actions (PEVA).

We trained a model to Predict Ego-centric Video from human Actions (PEVA) for Whole-Body-Conditioned Egocentric Video Prediction.

We train an autoregressive conditional diffusion transformer on Nymeria, a large-scale dataset pairing real-world egocentric video with body pose capture.

We include some samples here:Body Movement Actions Move Forward Rotate Left Rotate Right Left Hand Actions Move Left Hand Up Move Left Hand Down Move Left Hand Left Move Left Hand Right Right Hand Actions Move Right Hand Up Move Right Hand Down Move Right Hand Left Move Right Hand RightLong RolloutHere you…

6 months, 2 weeks назад @ bair.berkeley.edu
Defending against Prompt Injection with Structured Queries (StruQ) and Preference Optimization (SecAlign)
Defending against Prompt Injection with Structured Queries (StruQ) and Preference Optimization (SecAlign) Defending against Prompt Injection with Structured Queries (StruQ) and Preference Optimization (SecAlign)

Defending against Prompt Injection with Structured Queries (StruQ) and Preference Optimization (SecAlign)Recent advances in Large Language Models (LLMs) enable exciting LLM-integrated applications.

To mitigate the imminent prompt injection threat, we propose two fine-tuning-defenses, StruQ and SecAlign.

Prompt Injection Attack: CausesBelow is the threat model of prompt injection attacks.

Prompt injection threat model in LLM-integrated applicationsWe propose that prompt injection has two causes.

Below are resources to learn more and keep updated on prompt injection attacks and defenses.

9 months, 1 week назад @ bair.berkeley.edu
Repurposing Protein Folding Models for Generation with Latent Diffusion
Repurposing Protein Folding Models for Generation with Latent Diffusion Repurposing Protein Folding Models for Generation with Latent Diffusion

Repurposing Protein Folding Models for Generation with Latent DiffusionPLAID is a multimodal generative model that simultaneously generates protein 1D sequence and 3D structure, by learning the latent space of protein folding models.

In PLAID, we develop a method that learns to sample from the latent space of protein folding models to generate new proteins.

Unlike many previous protein structure generative models, PLAID addresses the multimodal co-generation problem setting: simultaneously generating both discrete sequence and continuous all-atom structural coordinates.

In this way, we can use structural understanding information in the weights of pretrained protein folding models for the p…

9 months, 1 week назад @ bair.berkeley.edu
Scaling Up Reinforcement Learning for Traffic Smoothing: A 100-AV Highway Deployment
Scaling Up Reinforcement Learning for Traffic Smoothing: A 100-AV Highway Deployment Scaling Up Reinforcement Learning for Traffic Smoothing: A 100-AV Highway Deployment

Scaling Up Reinforcement Learning for Traffic Smoothing: A 100-AV Highway DeploymentTraining Diffusion Models with Reinforcement LearningWe deployed 100 reinforcement learning (RL)-controlled cars into rush-hour highway traffic to smooth congestion and reduce fuel consumption for everyone.

The challenges of phantom jamsA stop-and-go wave moving backwards through highway traffic.

Smoothing behavior of RL AVs.

Overall, the steps towards deployment involved:Training in data-driven simulations: We used highway traffic data from I-24 to create a training environment with realistic wave dynamics, then validate the trained agent’s performance and robustness in a variety of new traffic scenarios.…

9 months, 3 weeks назад @ bair.berkeley.edu
AWS Machine Learning AWS Machine Learning
последний пост 18 часов назад
Securing Amazon Bedrock cross-Region inference: Geographic and global
Securing Amazon Bedrock cross-Region inference: Geographic and global Securing Amazon Bedrock cross-Region inference: Geographic and global

Amazon Bedrock provides two types of cross-Region inference profiles:Geographic cross-Region inference – Amazon Bedrock automatically selects the optimal Region within a defined geography (such as the US, EU, Australia, and Japan) to process your inference request.

Additionally, to confirm that the permission applies only to a specific Global cross-Region inference profile, you can use condition bedrock:InferenceProfileArn with the value of Global cross-Region inference profile ARN.

*" ] } } } ] }To restrict Geographic cross-Region inference for specific IAM roles or users, prevent assigning IAM policies with Geographic cross-Region inference permissions to specific IAM users or roles.

*" }…

18 часов назад @ aws.amazon.com
How Omada Health scaled patient care by fine-tuning Llama models on Amazon SageMaker AI
How Omada Health scaled patient care by fine-tuning Llama models on Amazon SageMaker AI How Omada Health scaled patient care by fine-tuning Llama models on Amazon SageMaker AI

In this post, we examine how Omada partnered with AWS and Meta to develop this healthcare-aligned AI solution using Llama models on Amazon SageMaker AI.

Amazon SageMaker Studio is used to launch a training job using Hugging Face estimators for fine-tuning Llama 3.1 8B model.

Omada centered their technical implementation around SageMaker AI for model training, fine-tuning, and deployment.

By using Llama models on SageMaker AI, Omada amplifies the humanity of health coaches and further enriches the member experience.

He works with customers to enable and accelerate their ML deployments on services such as Amazon SageMaker and Amazon EC2.

2 days назад @ aws.amazon.com
Crossmodal search with Amazon Nova Multimodal Embeddings
Crossmodal search with Amazon Nova Multimodal Embeddings Crossmodal search with Amazon Nova Multimodal Embeddings

In this post, we explore how Amazon Nova Multimodal Embeddings addresses the challenges of crossmodal search through a practical ecommerce use case.

The search problemTraditional approaches involve keyword-based search, text embeddings-based natural language search, or hybrid search and can’t process visual queries effectively, creating a gap between user intent and retrieval capabilities.

How Amazon Nova Multimodal Embeddings helpsAmazon Nova handles different types of search queries through the same model, which creates both new search capabilities and technical advantages.

Crossmodal search capabilitiesAs previously stated, Amazon Nova Multimodal Embeddings processes all supported modali…

4 days, 17 hours назад @ aws.amazon.com
Accelerating LLM inference with post-training weight and activation using AWQ and GPTQ on Amazon SageMaker AI
Accelerating LLM inference with post-training weight and activation using AWQ and GPTQ on Amazon SageMaker AI Accelerating LLM inference with post-training weight and activation using AWQ and GPTQ on Amazon SageMaker AI

Quick way to test ultra-low weight precision; helps gauge if 4-bit quantization is feasible before moving to more optimized schemes.

Using Amazon SageMaker AI for inference optimization and model quantizationIn this section, we cover how to implement quantization using Amazon SageMaker AI.

Step 1: Load model using HuggingFace transformersLoad the model weights without attaching them to an accelerator.

def quantize_model( args: argparse.Namespace ) -> None: try: ... # load model model = AutoModelForCausalLM.from_pretrained( args.model_id, torch_dtype="auto", device_map=None, trust_remote_code=True ) # load tokenizer tokenizer_or_processor = AutoTokenizer.from_pretrained( args.model_id, trust…

4 days, 23 hours назад @ aws.amazon.com
How Beekeeper optimized user personalization with Amazon Bedrock
How Beekeeper optimized user personalization with Amazon Bedrock How Beekeeper optimized user personalization with Amazon Bedrock

Quality is scored with a small synthetic test set and validated in production with user feedback (thumbs up/down and comments).

Solution overviewBeekeeper’s solution consists of two main phases: building a baseline leaderboard and personalizing with user feedback.

The coordinator fetches ranked model/prompt pairs and sends them with user feedback to a mutator, which returns personalized prompts.

However, they don’t want user feedback to fully change the behavior of their model through prompt injection in feedback.

By combining the benefits of synthetic data with user feedback, the solution is suitable even for smaller engineering teams.

5 days, 1 hour назад @ aws.amazon.com
Sentiment Analysis with Text and Audio Using AWS Generative AI Services: Approaches, Challenges, and Solutions
Sentiment Analysis with Text and Audio Using AWS Generative AI Services: Approaches, Challenges, and Solutions Sentiment Analysis with Text and Audio Using AWS Generative AI Services: Approaches, Challenges, and Solutions

Sentiment analysis has grown increasingly important in modern enterprises, providing insights into customer opinions, satisfaction levels, and potential frustrations.

We explore audio-based sentiment analysis in two stages:Stage 1 – Transcribe audio into text and perform sentiment analysis using LLMs– Transcribe audio into text and perform sentiment analysis using LLMs Stage 2 – Analyze sentiment directly from the audio signal using audio modelsSentiment analysis in textIn this section, we discuss the method of transcribing audio into text and performing sentiment analysis using LLMS.

Sentiment analysis in audioIn this section, we discuss the method of analyzing sentiment directly from the …

5 days, 1 hour назад @ aws.amazon.com
Architecting TrueLook’s AI-powered construction safety system on Amazon SageMaker AI
Architecting TrueLook’s AI-powered construction safety system on Amazon SageMaker AI Architecting TrueLook’s AI-powered construction safety system on Amazon SageMaker AI

To build, train, and deploy these models, TrueLook uses SageMaker AI, which provides managed infrastructure for the entire ML workflow.

TrueLook’s labeled image dataset moves through a training pipeline in three key stages: preprocessing (SageMaker Processing Job), training (SageMaker Training Job), and versioning with observability (SageMaker Model Registry).

SageMaker Training jobs executed the model training using built-in PyTorch containers and NVIDIA GPUs.

Operationalizing with SageMaker AIBy using SageMaker AI, TrueLook operationalized its multi-stage object detection workflow as a scalable, production-ready MLOps framework.

Implementing end-to-end object detection using SageMaker Pip…

5 days, 1 hour назад @ aws.amazon.com
Scaling medical content review at Flo Health using Amazon Bedrock (Part 1)
Scaling medical content review at Flo Health using Amazon Bedrock (Part 1) Scaling medical content review at Flo Health using Amazon Bedrock (Part 1)

This is why the team at Flo Health, the company behind the leading women’s health app Flo, is using generative AI to facilitate medical content accuracy at scale.

Solution overviewIn this section, we outline how the MACROS solution uses Amazon Bedrock and other AWS services to automate medical content review and revisions.

The UI support enables medical experts to directly see the content review statistics, interact with changes, and do manual adjustments.

Content review and revisionAt the core of MACROS lies its Content Review and Revision functionality with Amazon Bedrock foundation models.

Preliminary ResultsOur Proof of Concept delivered strong results across the critical success metric…

5 days, 23 hours назад @ aws.amazon.com
Detect and redact personally identifiable information using Amazon Bedrock Data Automation and Guardrails
Detect and redact personally identifiable information using Amazon Bedrock Data Automation and Guardrails Detect and redact personally identifiable information using Amazon Bedrock Data Automation and Guardrails

This post shows an automated PII detection and redaction solution using Amazon Bedrock Data Automation and Amazon Bedrock Guardrails through a use case of processing text and image content in high volumes of incoming emails and attachments.

Bootstrap the AWS account to use AWS CDK cdk bootstrap Users can now synthesize the CloudFormation template for this code.

ConclusionIn this post, we demonstrated how to automate the detection and redaction of PII across both text and image content using Amazon Bedrock Data Automation and Amazon Bedrock Guardrails.

By using Amazon Bedrock Data Automation and Amazon Bedrock Guardrails centralized redaction capabilities, organizations can boost data privac…

6 days, 1 hour назад @ aws.amazon.com
Speed meets scale: Load testing SageMakerAI endpoints with Observe.AI’s testing tool
Speed meets scale: Load testing SageMakerAI endpoints with Observe.AI’s testing tool Speed meets scale: Load testing SageMakerAI endpoints with Observe.AI’s testing tool

Select Sagemaker from the navigation pane and configure the test: SageMaker endpoint – Enter the name of the SageMaker endpoint from the SageMaker Unified Studio console here.

– Enter the AWS access key generated from AWS STS in the section “Generate your AWS credentials using AWS STS” above.

AWS secret key – Enter the AWS secret key generated from AWS STS in the section “Generate your AWS credentials using AWS STS” above.

– Enter the AWS secret key generated from AWS STS in the section “Generate your AWS credentials using AWS STS” above.

AWS session token – Enter the session token generated from AWS STS in the section “Generate your AWS credentials using AWS STS” above.

6 days, 1 hour назад @ aws.amazon.com
Migrate MLflow tracking servers to Amazon SageMaker AI with serverless MLflow
Migrate MLflow tracking servers to Amazon SageMaker AI with serverless MLflow Migrate MLflow tracking servers to Amazon SageMaker AI with serverless MLflow

You can apply the same approach to migrate existing SageMaker managed MLflow tracking servers to the new serverless MLflow capability on SageMaker.

Whichever environment you select must maintain connectivity to both your source tracking server and your target tracking server.

MLflow Export Import supports exports from both self-managed tracking servers and Amazon SageMaker MLflow tracking servers (from MLflow v2.16 onwards) to Amazon SageMaker Serverless MLflow.

To prepare for a successful migration:Verify the current MLflow version of your existing MLflow tracking server: mlflow --version Review the latest supported MLflow version in the Amazon SageMaker MLflow documentation.

For more info…

2 weeks, 2 days назад @ aws.amazon.com
Build an AI-powered website assistant with Amazon Bedrock
Build an AI-powered website assistant with Amazon Bedrock Build an AI-powered website assistant with Amazon Bedrock

This post demonstrates how to solve this challenge by building an AI-powered website assistant using Amazon Bedrock and Amazon Bedrock Knowledge Bases.

Amazon Bedrock managed LLMs – A large language model (LLM) from Amazon Bedrock generates AI-powered responses to user questions.

Model access in Amazon Bedrock for Amazon Titan and Amazon Nova Lite.

Create knowledge base and ingest website dataThe first step is to build a knowledge base to ingest data from a website and operational documents from an S3 bucket.

For Data source name, enter a name for your data source.

2 weeks, 2 days назад @ aws.amazon.com
Programmatically creating an IDP solution with Amazon Bedrock Data Automation
Programmatically creating an IDP solution with Amazon Bedrock Data Automation Programmatically creating an IDP solution with Amazon Bedrock Data Automation

Today, we explore how to programmatically create an IDP solution that uses Strands SDK, Amazon Bedrock AgentCore, Amazon Bedrock Knowledge Base, and Bedrock Data Automation (BDA).

Amazon Bedrock Data Automation can be used as a standalone feature or as a parser when setting up a knowledge base for Retrieval-Augmented Generation (RAG) workflows.

Amazon Bedrock AgentCore is a fully managed service that allows you to build and configure autonomous agents.

Here’s an overview of how you can setup Bedrock Knowledge Bases with data automation as a parser with Bedrock AgentCore.

With Amazon Bedrock Data Automation, we can enhance the RAG experience for more complex data formats including visual ric…

3 weeks назад @ aws.amazon.com
AI agent-driven browser automation for enterprise workflow management
AI agent-driven browser automation for enterprise workflow management AI agent-driven browser automation for enterprise workflow management

This workflow demonstrates the full capabilities of AI-powered browser automation, from initial navigation through complex decision-making to human-in-the-loop intervention.

Amazon Bedrock AgentCore Browser provides a secure, cloud-based browser that enables the AI agent (Amazon Nova Act and Strands agent in this case) to interact with websites.

ConclusionAI agent-driven browser automation represents a fundamental shift in how enterprises approach workflow management.

Veda Raman is a Sr Solutions Architect for Generative AI for Amazon Nova and Agentic AI at AWS.

She helps customers design and build Agentic AI solutions using Amazon Nova models and Bedrock AgentCore.

3 weeks назад @ aws.amazon.com
Agentic QA automation using Amazon Bedrock AgentCore Browser and Amazon Nova Act
Agentic QA automation using Amazon Bedrock AgentCore Browser and Amazon Nova Act Agentic QA automation using Amazon Bedrock AgentCore Browser and Amazon Nova Act

In this post, we explore how agentic QA automation addresses these challenges and walk through a practical example using Amazon Bedrock AgentCore Browser and Amazon Nova Act to automate testing for a sample retail application.

Benefits of agentic QA testingAgentic AI shifts QA testing from rule-based automation to intelligent, autonomous testing systems.

AgentCore Browser for large-scale agentic QA testingTo realize the potential of agentic AI testing at enterprise scale, organizations need robust infrastructure that can support intelligent, autonomous testing agents.

Agentic QA with the Amazon Nova Act SDKThe infrastructure capabilities of AgentCore Browser become truly powerful when combi…

3 weeks назад @ aws.amazon.com
NVIDIA
последний пост 21 час назад
Learn How NVIDIA cuOpt Accelerates Mixed Integer Optimization using Primal Heuristics
Learn How NVIDIA cuOpt Accelerates Mixed Integer Optimization using Primal Heuristics Learn How NVIDIA cuOpt Accelerates Mixed Integer Optimization using Primal Heuristics

NVIDIA cuOpt is a GPU-accelerated optimization engine designed to deliver fast, high-quality solutions for large, complex decision-making problems.

Accelerated primal heuristics for MIP solvers are algorithms that deliver high-quality, feasible solutions without exhaustively searching the entire solution space.

First, many real-world MIP problems are too large or time-sensitive for traditional solvers to find answers in time.

This indicates potential for further improvement, as well as the possibility of augmenting any existing solver with the GPU-accelerated primal heuristics detailed previously.

Get started with the MIP heuristics solverMIP heuristics offer fast and feasible solutions wit…

21 час назад @ developer.nvidia.com
CEOs of NVIDIA and Lilly Share ‘Blueprint for What Is Possible’ in AI and Drug Discovery
CEOs of NVIDIA and Lilly Share ‘Blueprint for What Is Possible’ in AI and Drug Discovery CEOs of NVIDIA and Lilly Share ‘Blueprint for What Is Possible’ in AI and Drug Discovery

NVIDIA and Lilly are putting together “a blueprint for what is possible in the future of drug discovery,” NVIDIA founder and CEO Jensen Huang told attendees at a fireside chat Monday with Dave Ricks, chair and CEO of Lilly.

This framework aims to enable experiments, data generation and AI model development to continuously inform and improve one another.

, developer of the Neo model family for co-folding and design across all biological molecules.

Max Jaderberg, president of Isomorphic , which is extending the capabilities of AlphaFold, the defining family of protein structure and interaction models.

Joshua Meier, CEO of Chai Discovery , which developed the Chai family of generative AI model…

21 час назад @ blogs.nvidia.com
AI’s Next Revolution: Multiply Labs Is Scaling Robotics-Driven Cell Therapy Biomanufacturing Labs
AI’s Next Revolution: Multiply Labs Is Scaling Robotics-Driven Cell Therapy Biomanufacturing Labs AI’s Next Revolution: Multiply Labs Is Scaling Robotics-Driven Cell Therapy Biomanufacturing Labs

Multiply Labs is doing for cell therapy labs what has already happened in the chip industry: It’s introducing robots to do the tedious, precision and hygienic work better, faster and cheaper.

“Next, I flew to Silicon Valley, and we started this at YCombinator.”San Francisco-based Multiply Labs, founded in 2016, today is automating cell therapy manufacturing with robots for leading companies, including Kyverna Therapeutics and Legend Biotech.

Multiply Labs offers end-to-end robotic systems that produce gene modified cell therapies at scale.

Robots within the controlled biomanufacturing clusters of Multiply Labs help ensure more hygienic and precision processes.

Simulating Cell Therapy Manufa…

2 days, 2 hours назад @ blogs.nvidia.com
NVIDIA Unveils Multi-Agent Intelligent Warehouse and Catalog Enrichment AI Blueprints to Power the Retail Pipeline
NVIDIA Unveils Multi-Agent Intelligent Warehouse and Catalog Enrichment AI Blueprints to Power the Retail Pipeline NVIDIA Unveils Multi-Agent Intelligent Warehouse and Catalog Enrichment AI Blueprints to Power the Retail Pipeline

New Multi-Agent Intelligent Warehouse (MAIW) and Retail Catalog Enrichment NVIDIA Blueprints are designed to turn this dynamic system into an advantage.

The NVIDIA MAIW blueprint delivers a synchronized AI system that sits above existing warehouse management systems, enterprise resource planning, robotics and IoT data, so teams gain real-time, explainable operational intelligence.

In addition, the Retail Catalog Enrichment Blueprint can create rich, on-brand marketing content by applying brand voice, tone and taxonomy instructions via prompts, alongside the product image and a target locale.

Global tech consulting firm Grid Dynamics has built a catalog enrichment and management system that …

5 days, 3 hours назад @ blogs.nvidia.com
Multi-Agent Warehouse AI Command Layer Enables Operational Excellence and Supply Chain Intelligence
Multi-Agent Warehouse AI Command Layer Enables Operational Excellence and Supply Chain Intelligence Multi-Agent Warehouse AI Command Layer Enables Operational Excellence and Supply Chain Intelligence

This post introduces the NVIDIA Multi-Agent Intelligent Warehouse (MAIW) Blueprint for the missing layer.

The solution: An AI command layerThe Multi-Agent Intelligent Warehouse delivers a unified AI command layer for modern warehouse operations, transforming fragmented systems, documents, and telemetry into real-time, actionable intelligence.

Intelligent document processingThe intelligent document processing pipeline uses NVIDIA NIM and multimodal foundation models with quality-based orchestration to deliver enterprise-grade accuracy at scale.

By instrumenting MAIW like any critical warehouse service, SRE and operations teams can monitor, debug, and improve the AI layer with confidence.

Lea…

5 days, 3 hours назад @ developer.nvidia.com
AI Copilot Keeps Berkeley’s X-Ray Particle Accelerator on Track
AI Copilot Keeps Berkeley’s X-Ray Particle Accelerator on Track AI Copilot Keeps Berkeley’s X-Ray Particle Accelerator on Track

Accelerator Assistant can help prepare and run a multistage physics experiment at a particle accelerator, reducing preparation effort by 100x.

In the rolling hills of Berkeley, California, an AI agent is supporting high-stakes physics experiments at the Advanced Light Source (ALS) particle accelerator.

Researchers at the Lawrence Berkeley National Laboratory ALS facility recently deployed the Accelerator Assistant, a large language model (LLM)-driven system to keep X-ray research on track.

And much can go wrong: the ALS control system has more than 230,000 process variables.

The work has already expanded beyond ALS as part of the DOE’s Genesys mission, with the framework being deployed acro…

6 days назад @ blogs.nvidia.com
Japan Science and Technology Agency Develops NVIDIA-Powered Moonshot Robot for Elderly Care
Japan Science and Technology Agency Develops NVIDIA-Powered Moonshot Robot for Elderly Care Japan Science and Technology Agency Develops NVIDIA-Powered Moonshot Robot for Elderly Care

Using NVIDIA Isaac Sim and RTX GPUs, humanoid robot research will automate cooking, cleaning, repositioning and other caregiving tasks.

In light of Japan’s rising elderly population, many of the research projects underway center on how robots can aid in senior care.

NVIDIA Architecture Powers On Moonshot RobotsNVIDIA technologies are integrated into every level of the Moonshot project’s senior care robots known as AI-Driven Robot for Embrace and Care, or AIREC.

Dry-AIREC robot, the larger and more mobile member of the Moonshot family, has two NVIDIA GPUs onboard.

Automating repositioning with a humanoid robot — while considering the elderly care patients’ personal states and bodily needs — …

6 days, 1 hour назад @ blogs.nvidia.com
More Ways to Play, More Games to Love — GeForce NOW Wraps CES With Linux Support, Fire TV App, Flight Stick Controls
More Ways to Play, More Games to Love — GeForce NOW Wraps CES With Linux Support, Fire TV App, Flight Stick Controls More Ways to Play, More Games to Love — GeForce NOW Wraps CES With Linux Support, Fire TV App, Flight Stick Controls

New AAA titles are streaming to the cloud, with easier automatic sign-in and six games joining the GeForce NOW library this week.

NVIDIA is wrapping up a big week at the CES trade show with a set of GeForce NOW announcements that are bringing more ways to play and more games to the cloud.

Leveling Up Where Gamers PlayA new native GeForce NOW app is launching in beta for Linux, starting with Ubuntu 24.04 and later.

Easier Sign-Ins, Faster Game TimeGeForce NOW now supports new games — and more ways to get into them faster.

Battle.net single sign-on recently joined the service, letting members jump straight into supported titles without juggling extra credentials each time.

6 days, 3 hours назад @ blogs.nvidia.com
Steel, Sensors and Silicon: How Caterpillar Is Bringing Edge AI to the Jobsite
Steel, Sensors and Silicon: How Caterpillar Is Bringing Edge AI to the Jobsite Steel, Sensors and Silicon: How Caterpillar Is Bringing Edge AI to the Jobsite

At CES this week, the future of technology was on display — and it wasn’t small.

That’s how Deepu Talla, vice president for robotics and edge AI at NVIDIA, wound up sharing the stage with Caterpillar for what is — when measured by sheer tonnage — the biggest demo at CES this year.

“Hey Cat, how do I get started?”A voice answered, generated by an AI system running directly on the machine.

At CES, that invisible layer streamed onto stage, paired with AI designed to help operators work more safely, efficiently and intuitively.

And the Cat 306 CR Mini Excavator, already recognized for precision and operator‑assist features, can be found at jobsites large and small across the world.

1 week назад @ blogs.nvidia.com
From Warehouse to Wallet: New State of AI in Retail and CPG Survey Uncovers How AI Is Rewiring Supply Chains and Customer Experiences
From Warehouse to Wallet: New State of AI in Retail and CPG Survey Uncovers How AI Is Rewiring Supply Chains and Customer Experiences From Warehouse to Wallet: New State of AI in Retail and CPG Survey Uncovers How AI Is Rewiring Supply Chains and Customer Experiences

AI agents are increasing the speed and efficiency of operations, while physical AI systems are helping streamline and automate warehouse and supply chain operations.

“Most retailers first started experimenting with AI using proprietary AI vendors,” said Jason Goldberg, chief commerce strategy officer of Publicis Groupe.

Agentic AI Makes Big Debut in RetailThe retail and CPG industry is piloting AI agents across lines of business.

The top pressure valve is using AI for supply chain operational efficiency and throughput, according to 51% of respondents.

Download the “State of AI in Retail and CPG: 2026 Trends” report for in-depth results and insights.

1 week назад @ blogs.nvidia.com
NVIDIA Brings GeForce RTX Gaming to More Devices With New GeForce NOW Apps for Linux PC and Amazon Fire TV
NVIDIA Brings GeForce RTX Gaming to More Devices With New GeForce NOW Apps for Linux PC and Amazon Fire TV NVIDIA Brings GeForce RTX Gaming to More Devices With New GeForce NOW Apps for Linux PC and Amazon Fire TV

Powered by GeForce RTX 5080-class performance on the NVIDIA Blackwell RTX platform, GeForce NOW Ultimate continues to raise the bar for PC gamers streaming from the cloud.

New this year, GeForce NOW is expanding that performance to more platforms than ever, headlined by a native Linux PC app and a new app for Amazon Fire TV sticks.

With rendering handled in the cloud, high-end PC gaming is possible on Linux operating systems, breathing new life into older devices.

This builds on existing TV support and helps make GeForce NOW the easiest way to bring high-performance PC gaming into the living room.

Take FlightGeForce NOW turns more devices into powerful cloud gaming rigs, and CES this year b…

1 week, 1 day назад @ blogs.nvidia.com
NVIDIA RTX Accelerates 4K AI Video Generation on PC With LTX-2 and ComfyUI Upgrades
NVIDIA RTX Accelerates 4K AI Video Generation on PC With LTX-2 and ComfyUI Upgrades NVIDIA RTX Accelerates 4K AI Video Generation on PC With LTX-2 and ComfyUI Upgrades

RTX Video Super Resolution integration in ComfyUI, accelerating 4K video generation.

A new video generation pipeline for generating 4K AI video using a 3D scene in Blender to precisely control outputs.

Once a video clip is generated, videos are upscaled to 4K in just seconds using the new RTX Video node in ComfyUI.

The video generation workflow will be available for download next month, with the newly released open weights of the LTX-2 Video Model and ComfyUI RTX updates available now.

Plug in to NVIDIA AI PC on Facebook, Instagram, TikTok and X — and stay informed by subscribing to the RTX AI PC newsletter.

1 week, 1 day назад @ blogs.nvidia.com
NVIDIA DLSS 4.5, Path Tracing and G-SYNC Pulsar Supercharge Gameplay With Enhanced Performance and Visuals
NVIDIA DLSS 4.5, Path Tracing and G-SYNC Pulsar Supercharge Gameplay With Enhanced Performance and Visuals NVIDIA DLSS 4.5, Path Tracing and G-SYNC Pulsar Supercharge Gameplay With Enhanced Performance and Visuals

NVIDIA DLSS 4.5 Will Power 4K 240Hz Path-Traced GamingNVIDIA DLSS 4.5 introduces Dynamic Multi Frame Generation and a new 6X Multi Frame Generation mode.

Dynamic Multi Frame Generation and 6X Multi-Frame Generation are expected to be available in spring of this year.

A second-generation transformer model for DLSS Super Resolution also arrives with NVIDIA DLSS 4.5, bringing state-of-the-art image quality to over 400 games and apps for all GeForce RTX GPUs.

The second-generation transformer is available to try now via the NVIDIA App for all GeForce RTX GPUs.

A new RTX Remix update — RTX Remix Logic — will be available later this month via the NVIDIA App.

1 week, 1 day назад @ blogs.nvidia.com
NVIDIA Rubin Platform, Open Models, Autonomous Driving: NVIDIA Presents Blueprint for the Future at CES
NVIDIA Rubin Platform, Open Models, Autonomous Driving: NVIDIA Presents Blueprint for the Future at CES NVIDIA Rubin Platform, Open Models, Autonomous Driving: NVIDIA Presents Blueprint for the Future at CES

NVIDIA founder and CEO Jensen Huang opened CES in Las Vegas with Rubin — NVIDIA’s first extreme-codesigned AI platform — plus open models for healthcare, robotics and autonomy, and a Mercedes-Benz CLA showcasing AI-defined driving.

Noting that 80% of startups are building on open models, Huang also emphasized the role of NVIDIA open models across every domain, trained on NVIDIA supercomputers, forming a global ecosystem of intelligence that developers and enterprises can build on.

Put it all together and the Rubin platform promises to dramatically accelerate AI innovation, delivering AI tokens at one-tenth the cost.

This is technology leadership.”Open Models for AllNVIDIA’s open models — tr…

1 week, 1 day назад @ blogs.nvidia.com
New Software and Model Optimizations Supercharge NVIDIA DGX Spark
New Software and Model Optimizations Supercharge NVIDIA DGX Spark New Software and Model Optimizations Supercharge NVIDIA DGX Spark

At CES 2026, the latest DGX Spark software release, combined with new model updates, and open-source libraries provide significant performance improvements for both DGX Spark and OEM GB10-based systems.

DGX Spark Llama.cpp performance improvementsA powerful desktop platform for creatorsWhile DGX Spark is an exceptional platform for AI developers, creators can also take advantage of its desktop-class capabilities.

DGX Spark and OEM GB10-based systems are now included in the program, with DGX Spark and partner systems currently in testing.

Access your DGX Spark from anywhere with NVIDIA BrevWith NVIDIA Brev, your DGX Spark is accessible from anywhere through a secure connection.

Using DGX Spa…

1 week, 1 day назад @ developer.nvidia.com
Facebook
последний пост 3 weeks, 5 days назад
DrP: Meta’s Root Cause Analysis Platform at Scale
DrP: Meta’s Root Cause Analysis Platform at Scale DrP: Meta’s Root Cause Analysis Platform at Scale

DrP’s key components include:Expressive SDK : The DrP SDK allows engineers to codify investigation workflows into analyzers.

Post-processing system : After an investigation, the post-processing system can take automated actions based on the analysis results.

Bootstrap code : The DrP SDK provides bootstrap code to create a template analyzer with pre-populated boilerplate code.

Data access and analysis : The SDK includes libraries for data access and analysis, such as dimension analysis and time series correlation.

This provides immediate analysis results to on-call engineers.

3 weeks, 5 days назад @ engineering.fb.com
How AI Is Transforming the Adoption of Secure-by-Default Mobile Frameworks
How AI Is Transforming the Adoption of Secure-by-Default Mobile Frameworks How AI Is Transforming the Adoption of Secure-by-Default Mobile Frameworks

Generative AI and automation accelerate the adoption of secure frameworks at scale, enabling consistent security enforcement and efficient migration across Meta’s vast codebase.

How We Design Secure-by-Default Frameworks at MetaDesigning secure-by-default frameworks for use by a large number of developers shipping vastly different features across multiple apps is an interesting challenge.

There shouldn’t be one security framework that covers all security issues, and not every security issue is general enough to deserve its own framework.

Now that we’ve looked at the design philosophy behind our frameworks, let’s look at one of our most widely used Android security frameworks, SecureLinkLaun…

1 month назад @ engineering.fb.com
Zoomer: Powering AI Performance at Meta’s Scale Through Intelligent Debugging and Optimization
Zoomer: Powering AI Performance at Meta’s Scale Through Intelligent Debugging and Optimization Zoomer: Powering AI Performance at Meta’s Scale Through Intelligent Debugging and Optimization

Zoomer has delivered training time reductions, and significant QPS improvements, making it the de-facto tool for AI performance optimization across Meta’s entire AI infrastructure.

Zoomer is Meta’s automated, one-stop-shop platform for performance profiling, debugging, analysis, and optimization of AI training and inference workloads.

AI Performance Optimization Using ZoomerZoomer is an automated debugging and optimization platform that works across all of our AI model types (ads recommendations, GenAI, computer vision, etc.)

Memory Analysis : Comprehensive analysis of GPU memory usage patterns, allocation tracking, and leak detection.

Realtime Memory Profiling : GPU memory allocation track…

1 month, 3 weeks назад @ engineering.fb.com
Open Source Is Good for the Environment
Open Source Is Good for the Environment Open Source Is Good for the Environment

But have you heard about open hardware?

And did you know open source can have a positive impact on the environment?

On this episode of the Meta Tech Podcast, Pascal Hartig sits down with Dharmesh and Lisa to talk about all things open hardware, and Meta’s biggest announcements from the 2025 Open Compute Project (OCP) Summit – including a new open methodology for leveraging AI to understand Scope 3 emissions.

You’ll also hear how AI and open hardware are helping Meta push to achieve net zero emissions in 2030, including how AI is being used to develop new concrete mixes for data center construction.

And if you’re interested in learning more about career opportunities at Meta visit the Meta C…

2 months назад @ engineering.fb.com
Meta’s Generative Ads Model (GEM): The Central Brain Accelerating Ads Recommendation AI Innovation
Meta’s Generative Ads Model (GEM): The Central Brain Accelerating Ads Recommendation AI Innovation Meta’s Generative Ads Model (GEM): The Central Brain Accelerating Ads Recommendation AI Innovation

We’re sharing details about Meta’s Generative Ads Recommendation Model (GEM), a new foundation model that delivers increased ad performance and advertiser ROI by enhancing other ads recommendation models’ ability to serve relevant ads.

GEM propagates its learnings, leveraging a suite of post-training techniques across the entire ads model fleet, enabling a paradigm shift in Meta’s Ads Recommendation system.

GEM leverages enhanced training scalability that efficiently utilizes thousands of GPUs for building and iterating an LLM-scale ads foundation model.

The Generative Ads Recommendation Model (GEM) is Meta’s most advanced ads foundation model, built on an LLM-inspired paradigm and trained …

2 months назад @ engineering.fb.com
Scaling LLM Inference: Innovations in Tensor Parallelism, Context Parallelism, and Expert Parallelism
Scaling LLM Inference: Innovations in Tensor Parallelism, Context Parallelism, and Expert Parallelism Scaling LLM Inference: Innovations in Tensor Parallelism, Context Parallelism, and Expert Parallelism

At Meta, we are constantly pushing the boundaries of LLM inference systems to power applications such as the Meta AI App.

These metrics highlight the distinct computational demands of LLM inference: Prefill is compute-intensive, while decoding is memory bandwidth-intensive.

Communication: Communication latency increases when parallelizing across multiple hosts.

In EP-based inference, we utilize a two-shot, all-to-all communication pattern to exchange tokens between data parallelism and expert parallelism ranks based on routing.

We are committed to continuous innovation to ensure efficient and scalable LLM inference for millions of users worldwide.

2 months, 4 weeks назад @ engineering.fb.com
How Meta Is Leveraging AI To Improve the Quality of Scope 3 Emission Estimates for IT Hardware
How Meta Is Leveraging AI To Improve the Quality of Scope 3 Emission Estimates for IT Hardware How Meta Is Leveraging AI To Improve the Quality of Scope 3 Emission Estimates for IT Hardware

We leveraged AI to help us improve this database and understand our Scope 3 emissions associated with IT hardware by:Identifying similar components and applying existing PCFs to similar components that lack these carbon estimates.

Understanding the carbon footprint of IT racks and applying generative AI (GenAI) as a categorization algorithm to create a new and standard taxonomy .

If these similar components are not identified their carbon footprint estimates will remain at a lower data quality.

These similar components can be mapped to a representative proxy PCF, allowing us to use high-quality PCF data in similar components.

For example, we can scale the carbon footprint calculation for a …

3 months назад @ engineering.fb.com
OCP Summit 2025: The Open Future of Networking Hardware for AI
OCP Summit 2025: The Open Future of Networking Hardware for AI OCP Summit 2025: The Open Future of Networking Hardware for AI

At Open Compute Project Summit (OCP) 2025, we’re sharing details about the direction of next-generation network fabrics for our AI training clusters.

At Meta, we believe that open hardware is a catalyst for innovation — especially as data center infrastructure increasingly supports new and emerging AI technologies.

Open hardware plays a crucial role in enabling disaggregation, allowing us to break down traditional data center technologies into their core components.

Today, through OCP, we continue to advance open network technologies for the next generation of AI applications.

Ethernet for Scale-Up Networking in OCP: Meta’s Industry LeadershipAt Meta, we recognize that the future of AI and …

3 months назад @ engineering.fb.com
LLMs Are the Key to Mutation Testing and Better Compliance
LLMs Are the Key to Mutation Testing and Better Compliance LLMs Are the Key to Mutation Testing and Better Compliance

By leveraging LLMs we’ve been able to overcome the barriers that have prevented mutation testing from being efficiently deployed at scale.

Our presentations shared insights into how we’ve used LLMs to solve the major barriers that have prevented mutation testing at scale and highlighted new areas in automated software testing where LLMs can have a significant impact.

Mutation Testing Isn’t ScalableTraditional mutation testing generates a very large number of mutants, making it computationally expensive and difficult to scale to large industrial codebases.

Mutation Testing Requires a Lot of Computational ResourcesMutation testing is costly in terms of computational resources and developer ef…

3 months, 2 weeks назад @ engineering.fb.com
AssetGen: Generating 3D Worlds With AI
AssetGen: Generating 3D Worlds With AI AssetGen: Generating 3D Worlds With AI

Imagine being able to use AI to create 3D virtual worlds using prompts as easily as you can generate images.

In his keynote, Mark Zuckerberg shared his vision of a future where anyone can create virtual worlds using AI-powered tools like the ones available in the upcoming Meta Horizon Studio.

But AI is already making it easier than ever to create 3D assets.

On this episode of the Meta Tech Podcast, Pascal Hartig is joined by Mahima and Rakesh from Meta’s XR Tech team to discuss AssetGen, a new foundation model for 3D assets.

They talk about how they built and trained AssetGen, the important role LLMs have to play in the future of VR, and how they’re tackling the ambitious goal of generating…

3 months, 2 weeks назад @ engineering.fb.com
Meta’s Infrastructure Evolution and the Advent of AI
Meta’s Infrastructure Evolution and the Advent of AI Meta’s Infrastructure Evolution and the Advent of AI

As our user base grew globally, we scaled beyond single data center buildings and into data center regions consisting of multiple buildings.

Enter AI Workloads (2020)While we were navigating the challenges of scaling, we were also seeing glimpses of how AI workloads would impact our infrastructure.

To build out our AI infrastructure, we’ve leveraged solutions from partners like AMD and NVIDIA as well as our own custom silicon.

Constructing Prometheus has been a monumental engineering feat, with infrastructure spanning five or more data center buildings in a single data center region.

We are still early in the evolution and adoption of AI workloads.

3 months, 2 weeks назад @ engineering.fb.com
Networking at the Heart of AI — @Scale: Networking 2025 Recap
Networking at the Heart of AI — @Scale: Networking 2025 Recap Networking at the Heart of AI — @Scale: Networking 2025 Recap

AI is everywhere and, as network engineers, we are right in the thick of it: building the network infrastructure for AI.

Setting Context: Rapid Changes and EvolutionGiven AI continues to drive so much innovation in networking and general infrastructure, we once again focused @Scale: Networking on AI networking, sharing the new insights and progress in the field.

The Models and the Primary AI Workloads Are Rapidly Evolving.

More from @Scale:Networking 2025Please visit the @Scale YouTube channel to check out all the talks from this year’s Networking @Scale.

We look forward to what promises to be another rapid year of network and AI innovation that we’ll cover at the next @Scale: Networking in…

3 months, 2 weeks назад @ engineering.fb.com
A New Ranking Framework for Better Notification Quality on Instagram
A New Ranking Framework for Better Notification Quality on Instagram A New Ranking Framework for Better Notification Quality on Instagram

We’ve introduced a diversity-aware notification ranking framework to reduce uniformity and deliver a more varied and engaging mix of notifications.

Instagram leverages machine learning (ML) models to decide who should get a notification, when to send it, and what content to include.

To tackle this, we’ve introduced a diversity-aware notification ranking framework that helps deliver more diverse, better curated, and less repetitive notifications.

Introducing Instagram’s Diversity-Aware Notification Ranking FrameworkInstagram’s diversity-aware notification ranking framework is designed to enhance the notification experience by balancing the predicted potential for user engagement with the nee…

4 months, 2 weeks назад @ engineering.fb.com
Federation Platform and Privacy Waves: How Meta distributes compliance-related tasks at scale
Federation Platform and Privacy Waves: How Meta distributes compliance-related tasks at scale Federation Platform and Privacy Waves: How Meta distributes compliance-related tasks at scale

We’re exploring Meta’s Federation Platform, a scalable set of tools for managing compliance-related tasks, along with Privacy Waves, our method for batching these tasks and ensuring accountability.

To facilitate this, we developed the Federation Platform and Privacy Waves program:The Federation Platform breaks down large compliance-related initiatives into smaller, manageable workstreams.

Internal surveys reveal significantly higher positive sentiment for Privacy Waves tasks compared to ad-hoc tasks.

Step 6: Reporting and recognitionThe centralized distribution of tasks via Federation Platform and Privacy Waves streamline operational effectiveness and verification.

Expansions for the Federa…

5 months назад @ engineering.fb.com
Diff Risk Score: AI-driven risk-aware software development
Diff Risk Score: AI-driven risk-aware software development Diff Risk Score: AI-driven risk-aware software development

Built on a fine-tuned Llama LLM, DRS evaluates code changes and metadata to produce a risk score and highlight potentially risky code snippets.

Production risk was one of the areas we tackled first.

The demand to build such features also led us to build the Risk Awareness Platform to provide risk analysis APIs and tool integrations.

We believe code risk can play a significant role in improving this tradeoff, so we will build more risk-aware features while improving their quality.

While code changes cause the plurality of SEVs at Meta, configuration changes are another large category.

5 months, 1 week назад @ engineering.fb.com
Uber Engineering
последний пост None
neptune.ai neptune.ai
последний пост 1 month, 1 week назад
We are joining OpenAI
We are joining OpenAI We are joining OpenAI

Piotr Niedźwiedź, CEO/CTO and founder of neptune.aiI’m excited to share that we’ve entered into a definitive agreement to be acquired by OpenAI, subject to closing conditions.

We are thrilled to join the OpenAI team and help their AI researchers build better models faster.

Neptune is a metrics dashboard company.”We’ve worked closely with OpenAI to create the metrics dashboard that helps teams building foundation models.

Our future with OpenAINeptune will join OpenAI and continue to support AI researchers with tools to monitor, debug, and evaluate frontier models.

We are looking forward to working with top AI researchers and supporting OpenAI’s mission of ensuring that AGI benefits all of hu…

1 month, 1 week назад @ neptune.ai
Synthetic Data for LLM Training
Synthetic Data for LLM Training Synthetic Data for LLM Training

For instance, financial data is highly sensitive and protected by very strict regulations, and synthetic data mimics the real data distribution without revealing customer information.

Read more about how leading foundation model teams curate their training data and other topics in the State of Foundation Model Training Report 2025.

Choosing the right synthetic data generation technique depends on the type of data and its complexity.

Synthetic tabular data generation is a promising direction to overcome these challenges by learning the distribution of the tabular data.

Post-processingAs the distribution of tabular data is highly complex, it makes the synthetic tabular data generation very ch…

2 months назад @ neptune.ai
What are LLM Embeddings: All you Need to Know
What are LLM Embeddings: All you Need to Know What are LLM Embeddings: All you Need to Know

TL;DR LLM embeddings are the numerical, vector representations of text that Large Language Models (LLMs) use to process information.

Unlike their predecessor word embeddings, LLM embeddings are context-aware and dynamically change to capture semantic and syntactic relationships based on the surrounding text.

What are the applications of LLM embeddings?

Word EmbeddingsSparse Word Embeddings One-Hot Vectors 1970s TF-IDF1980s Co-Occurrence MatrixStatic Word Embeddings Word2Vec 2013 GloVe 2014Contextualized word embeddings ELMo 2018 GPT-1 2018 BERT 2018 LLAMA 2023 DeepSeek-V1 2023 GPT-4 2023Static word embeddingsStatic word embeddings, such as word2vec in 2013, marked a significant development.…

2 months, 1 week назад @ neptune.ai
Detecting and Fixing ‘Dead Neurons’ in Foundation Models
Detecting and Fixing ‘Dead Neurons’ in Foundation Models Detecting and Fixing ‘Dead Neurons’ in Foundation Models

TL;DR Dead neurons silently waste compute and reduce effective model capacity in foundation models.

Dead neurons’ impactRecent studies into dead neurons in the context of foundation models show interesting, albeit worrying, results.

These large reported fractions of dead neurons in foundation models are a concern from a computational perspective.

Before we move on to discuss how to detect and fix dead neurons, let’s touch upon an important distinction between dead neurons and vanishing gradients.

Further reading How to Monitor, Diagnose, and Solve Gradient Issues in Foundation Models Read moreVisualizing activation distributionsIs your foundation model suffering from dead neurons?

2 months, 2 weeks назад @ neptune.ai
Part 2: Instruction Fine-Tuning: Evaluation and Advanced Techniques for Efficient Training
Part 2: Instruction Fine-Tuning: Evaluation and Advanced Techniques for Efficient Training Part 2: Instruction Fine-Tuning: Evaluation and Advanced Techniques for Efficient Training

In the first part of this series, we covered the fundamentals of instruction fine-tuning (IFT).

def calculate_irs(instruction, output, reference_model): evaluation_prompt = f""" Instruction: {instruction} Model Output: {output} Rate how well the output follows the instruction on these criteria: 1.

| SourceHINT addresses a computational inefficiency in standard instruction fine-tuning: repeatedly reprocessing the same task instruction with every input example.

Read more about foundation model training infrastructure and other topics in Neptune’s 2025 State of Foundation Model Training Report.

First, during initial instruction fine-tuning across multiple diverse tasks, the model learns genera…

2 months, 3 weeks назад @ neptune.ai
How to Optimize LLM Inference
How to Optimize LLM Inference How to Optimize LLM Inference

Large Language Model (LLM) inference at scale is challenging as it involves transferring massive amounts of model parameters and data and performing computations on large tensors.

In the following, we’ll use the Llama model family architecture as a specific example to understand the LLM workload at inference.

For a far more detailed analysis of the LLM workload at inference, see the chapter All About Transformer Inference in the book How to Scale Your Model, published by Google DeepMind.

See also How to Run LLMs Locally Read moreA quick primer on hardware for LLM inferenceA typical LLM inference cluster consists of several nodes, each with a multi-core CPU and multiple accelerator devices, …

3 months назад @ neptune.ai
A Researcher’s Guide to LLM Grounding
A Researcher’s Guide to LLM Grounding A Researcher’s Guide to LLM Grounding

In this article, we’ll explore the fundamental concepts of LLM grounding as well as strategies for optimally grounding models.

What is LLM grounding?

LLM grounding is analogous.

If relevant knowledge cannot be inferred from the data, then LLM grounding cannot yield more relevant responses.

When grounding LLMs using RAG, consider retaining only a few of the top hits (i.e., top-k) for your retrieval queries.

3 months, 2 weeks назад @ neptune.ai
Instruction Fine-Tuning: Fundamentals, Architecture Modifications, and Loss Functions
Instruction Fine-Tuning: Fundamentals, Architecture Modifications, and Loss Functions Instruction Fine-Tuning: Fundamentals, Architecture Modifications, and Loss Functions

TL;DR Instruction fine-tuning (IFT) refines pre-trained large language models (LLMs) to follow specific task instructions by training on prompt-response pairs.

Instruction fine-tuning in a nutshellIFT tailors LLMs to follow user instructions by bridging their inherent next-word prediction with human-defined objectives.

Related LLM Fine-Tuning and Model Selection Using Neptune and Transformers Read moreParameter-efficient instruction fine-tuningWhile major foundation models like GPT-4 or Llama-2 undergo full parameter instruction fine-tuning during development, parameter-efficient fine-tuning (PEFT) methods have become widely adopted for instruction fine-tuning since the LoRA paper was publi…

3 months, 4 weeks назад @ neptune.ai
Understanding Prompt Injection: Risks, Methods, and Defense Measures
Understanding Prompt Injection: Risks, Methods, and Defense Measures Understanding Prompt Injection: Risks, Methods, and Defense Measures

Prompt injection 101: When prompts go rogueThe term ‘Prompt Injection’ comes from SQL injection attacks.

There is another claim of the independent discovery of prompt injection attacks, which suggests that Riley Goodside publicly exhibited a prompt injection in a tweet back in September 2022.

The indirect prompt injection attacks are classified into active, passive, user-driven and virtual prompt attacks.

Virtual prompt injection attacksThis injection type is closely related to passive injection attacks previously described.

Prompt injection: current challenges & lessons learnedThe arms race between prompt injection attacks and defenses is a challenge for researchers, developers, and users.

5 months, 1 week назад @ neptune.ai
SabiYarn: Advancing Low-Resource Languages With Multitask NLP Pre-Training [Paper Reflections]
SabiYarn: Advancing Low-Resource Languages With Multitask NLP Pre-Training [Paper Reflections] SabiYarn: Advancing Low-Resource Languages With Multitask NLP Pre-Training [Paper Reflections]

This simple idea avoids computing loss on input prompt tokens the model already knows.

Prompt tokens are (too) expensive in low-resource settingsDuring pre-training, LLMs are trained in causal language modeling through a next-token prediction task.

=> Mo fẹ́ràn ìrẹsì,” the model is trained to predict every token, from the prompt to the actual answer:Step Prompt Next token 1 Translate English Static prompt 2 Translate English to Static prompt 3 Translate English to Yoruba: Static prompt 4 Translate English to Yoruba: I 5 Translate English to Yoruba: I love 6 Translate English to Yoruba: I love rice.

This is straightforward to implement in PyTorch by masking out the prompt tokens in the label …

5 months, 2 weeks назад @ neptune.ai
How to Monitor, Diagnose, and Solve Gradient Issues in Foundation Models
How to Monitor, Diagnose, and Solve Gradient Issues in Foundation Models How to Monitor, Diagnose, and Solve Gradient Issues in Foundation Models

What gradient issues occur during foundation model training?

During training, gradient descent updates model parameters by computing the gradients of the loss function via forward and backward passes.

The green line corresponds to a learning rate of 10, while the orange line has a learning rate of 0.1.

The gradient norm for the orange line with LR = 0.1 is very high in the first steps, while the gradient norm of the green line with LR = 10 diverges to NaN after a few steps.

Techniques for gradient stabilizationMonitoring gradient norms and training loss provides insights into the learning dynamics of the foundation models.

6 months, 2 weeks назад @ neptune.ai
STUN: Structured-Then-Unstructured Pruning for Scalable MoE Pruning [Paper Reflection]
STUN: Structured-Then-Unstructured Pruning for Scalable MoE Pruning [Paper Reflection] STUN: Structured-Then-Unstructured Pruning for Scalable MoE Pruning [Paper Reflection]

Unstructured pruning removes individual weights, while structured pruning removes entire model components.

In the context of MoEs, as expert structures from training MoEs correspond to such patterns, pruning experts is a natural fit for structured pruning.

Thus, structured pruning does not significantly decrease kurtosis, leaving plenty of margin for unstructured pruning.

Since structured pruning primarily reduces architectural redundancy rather than reshaping the underlying weight distribution, our two-phase approach—leveraging unstructured pruning after structured pruning—outperforms unstructured-only pruning.

Since STUN does not make any assumption about base MoE models, it is generaliza…

7 months, 1 week назад @ neptune.ai
Evaluating RAG Pipelines
Evaluating RAG Pipelines Evaluating RAG Pipelines

Related Building LLM Applications With Vector Databases Read moreDimensions of RAG evaluationEvaluating a RAG pipeline means assessing its behavior across three dimensions:1.

The evaluation of the RAG pipeline is a multi-step process, starting with creating an evaluation dataset, then evaluating the individual components (retriever, generator, etc.

Curating an evaluation datasetThe first step in the RAG evaluation process is the creation of a ground truth dataset.

MAP considers both the presence and rank of relevant chunks but fails to consider the relative position of relevant chunks.

However, not all retrieved chunks are equally relevant and sometimes, the most relevant chunks might not b…

8 months назад @ neptune.ai
▶️ YouTube
Yannic Kilcher Yannic Kilcher
последний пост 2 weeks, 3 days назад
Traditional X-Mas Stream
Traditional X-Mas Stream Traditional X-Mas Stream

Letsgooo

2 weeks, 3 days назад @ youtube.com
Traditional Holiday Live Stream
Traditional Holiday Live Stream Traditional Holiday Live Stream

https://ykilcher.com/discord Links:

TabNine Code Completion (Referral): http://bit.ly/tabnine-yannick

YouTube: https://www.youtube.com/c/yannickilcher

Twitter: https://twitter.com/ykilcher

Discord: https://discord.gg/4H8xxDF

BitChute: https://www.bitchute.com/channel/yannic-kilcher

Minds: https://www.minds.com/ykilcher

Parler: https://parler.com/profile/YannicKilcher

LinkedIn: https://www.linkedin.com/in/yannic-kilcher-488534136/

BiliBili: https://space.bilibili.com/1824646584 If you want to support me, the best thing to do is to share out the content :) If you want to support me financially (completely optional and voluntary, but a lot of people have asked for this):

SubscribeStar: https:/…

2 weeks, 3 days назад @ youtube.com
TiDAR: Think in Diffusion, Talk in Autoregression (Paper Analysis)
TiDAR: Think in Diffusion, Talk in Autoregression (Paper Analysis) TiDAR: Think in Diffusion, Talk in Autoregression (Paper Analysis)

Paper: https://arxiv.org/abs/2511.08923 Abstract:

Diffusion language models hold the promise of fast parallel generation, while autoregressive (AR) models typically excel in quality due to their causal structure aligning naturally with language modeling. This raises a fundamental question: can we achieve a synergy with high throughput, higher GPU utilization, and AR level quality? Existing methods fail to effectively balance these two aspects, either prioritizing AR using a weaker model for sequential drafting (speculative decoding), leading to lower drafting efficiency, or using some form of left-to-right (AR-like) decoding logic for diffusion, which still suffers from quality degradation …

2 weeks, 4 days назад @ youtube.com
Titans: Learning to Memorize at Test Time (Paper Analysis)
Titans: Learning to Memorize at Test Time (Paper Analysis) Titans: Learning to Memorize at Test Time (Paper Analysis)

Paper: https://arxiv.org/abs/2501.00663 Abstract:

Over more than a decade there has been an extensive research effort on how to effectively utilize recurrent models and attention. While recurrent models aim to compress the data into a fixed-size memory (called hidden state), attention allows attending to the entire context window, capturing the direct dependencies of all tokens. This more accurate modeling of dependencies, however, comes with a quadratic cost, limiting the model to a fixed-length context. We present a new neural long-term memory module that learns to memorize historical context and helps attention to attend to the current context while utilizing long past information. We sh…

1 month назад @ youtube.com
[Paper Analysis] The Free Transformer (and some Variational Autoencoder stuff)
[Paper Analysis] The Free Transformer (and some Variational Autoencoder stuff) [Paper Analysis] The Free Transformer (and some Variational Autoencoder stuff)

https://arxiv.org/abs/2510.17558 Abstract:

We propose an extension of the decoder Transformer that conditions its generative process on random latent variables which are learned without supervision thanks to a variational procedure. Experimental evaluations show that allowing such a conditioning translates into substantial improvements on downstream tasks. Author: François Fleuret Links:

Homepage: https://ykilcher.com

Merch: https://ykilcher.com/merch

YouTube: https://www.youtube.com/c/yannickilcher

Twitter: https://twitter.com/ykilcher

Discord: https://ykilcher.com/discord

LinkedIn: https://www.linkedin.com/in/ykilcher If you want to support me, the best thing to do is to share out the con…

2 months, 2 weeks назад @ youtube.com
[Video Response] What Cloudflare's code mode misses about MCP and tool calling
[Video Response] What Cloudflare's code mode misses about MCP and tool calling [Video Response] What Cloudflare's code mode misses about MCP and tool calling

Theo's Video: https://www.youtube.com/watch?v=bAYZjVAodoo

Cloudflare article: https://blog.cloudflare.com/code-mode/ Links:

Homepage: https://ykilcher.com

Merch: https://ykilcher.com/merch

YouTube: https://www.youtube.com/c/yannickilcher

Twitter: https://twitter.com/ykilcher

Discord: https://ykilcher.com/discord

LinkedIn: https://www.linkedin.com/in/ykilcher If you want to support me, the best thing to do is to share out the content :) If you want to support me financially (completely optional and voluntary, but a lot of people have asked for this):

SubscribeStar: https://www.subscribestar.com/yannickilcher

Patreon: https://www.patreon.com/yannickilcher

Bitcoin (BTC): bc1q49lsw3q325tr58ygf8…

2 months, 3 weeks назад @ youtube.com
[Paper Analysis] On the Theoretical Limitations of Embedding-Based Retrieval (Warning: Rant)
[Paper Analysis] On the Theoretical Limitations of Embedding-Based Retrieval (Warning: Rant) [Paper Analysis] On the Theoretical Limitations of Embedding-Based Retrieval (Warning: Rant)

Paper: https://arxiv.org/abs/2508.21038 Abstract:

Vector embeddings have been tasked with an ever-increasing set of retrieval tasks over the years, with a nascent rise in using them for reasoning, instruction-following, coding, and more. These new benchmarks push embeddings to work for any query and any notion of relevance that could be given. While prior works have pointed out theoretical limitations of vector embeddings, there is a common assumption that these difficulties are exclusively due to unrealistic queries, and those that are not can be overcome with better training data and larger models. In this work, we demonstrate that we may encounter these theoretical limitations in realist…

3 months назад @ youtube.com
AGI is not coming!
AGI is not coming! AGI is not coming!

jack Morris's investigation into GPT-OSS training data https://x.com/jxmnop/status/1953899426075816164?t=3YRhVQDwQLk2gouTSACoqA&s=09

5 months, 1 week назад @ youtube.com
Context Rot: How Increasing Input Tokens Impacts LLM Performance (Paper Analysis)
Context Rot: How Increasing Input Tokens Impacts LLM Performance (Paper Analysis) Context Rot: How Increasing Input Tokens Impacts LLM Performance (Paper Analysis)

Paper: https://research.trychroma.com/context-rot Abstract:

Large Language Models (LLMs) are typically presumed to process context uniformly—that is, the model should handle the 10,000th token just as reliably as the 100th. However, in practice, this assumption does not hold. We observe that model performance varies significantly as input length changes, even on simple tasks.

In this report, we evaluate 18 LLMs, including the state-of-the-art GPT-4.1, Claude 4, Gemini 2.5, and Qwen3 models. Our results reveal that models do not use their context uniformly; instead, their performance grows increasingly unreliable as input length grows. Authors: Kelly Hong, Anton Troynikov, Jeff Huber Links:

5 months, 3 weeks назад @ youtube.com
Energy-Based Transformers are Scalable Learners and Thinkers (Paper Review)
Energy-Based Transformers are Scalable Learners and Thinkers (Paper Review) Energy-Based Transformers are Scalable Learners and Thinkers (Paper Review)

Paper: https://arxiv.org/abs/2507.02092

Code: https://github.com/alexiglad/EBT

Website: https://energy-based-transformers.github.io/ Abstract:

Inference-time computation techniques, analogous to human System 2 Thinking, have recently become popular for improving model performances. However, most existing approaches suffer from several limitations: they are modality-specific (e.g., working only in text), problem-specific (e.g., verifiable domains like math and coding), or require additional supervision/training on top of unsupervised pretraining (e.g., verifiers or verifiable rewards). In this paper, we ask the question "Is it possible to generalize these System 2 Thinking approaches, and de…

5 months, 4 weeks назад @ youtube.com
On the Biology of a Large Language Model (Part 2)
On the Biology of a Large Language Model (Part 2) On the Biology of a Large Language Model (Part 2)

An in-depth look at Anthropic's Transformer Circuit Blog Post

Part 1 here: https://youtu.be/mU3g2YPKlsA

Discord here: https;//ykilcher.com/discord https://transformer-circuits.pub/2025/attribution-graphs/biology.html Abstract:

We investigate the internal mechanisms used by Claude 3.5 Haiku — Anthropic's lightweight production model — in a variety of contexts, using our circuit tracing methodology. Authors:

Jack Lindsey†, Wes Gurnee*, Emmanuel Ameisen*, Brian Chen*, Adam Pearce*, Nicholas L. Turner*, Craig Citro*,

David Abrahams, Shan Carter, Basil Hosmer, Jonathan Marcus, Michael Sklar, Adly Templeton,

Trenton Bricken, Callum McDougall◊, Hoagy Cunningham, Thomas Henighan, Adam Jermyn, Andy …

8 months, 2 weeks назад @ youtube.com
On the Biology of a Large Language Model (Part 1)
On the Biology of a Large Language Model (Part 1) On the Biology of a Large Language Model (Part 1)

An in-depth look at Anthropic's Transformer Circuit Blog Post https://transformer-circuits.pub/2025/attribution-graphs/biology.html Abstract:

We investigate the internal mechanisms used by Claude 3.5 Haiku — Anthropic's lightweight production model — in a variety of contexts, using our circuit tracing methodology. Authors:

Jack Lindsey†, Wes Gurnee*, Emmanuel Ameisen*, Brian Chen*, Adam Pearce*, Nicholas L. Turner*, Craig Citro*,

David Abrahams, Shan Carter, Basil Hosmer, Jonathan Marcus, Michael Sklar, Adly Templeton,

Trenton Bricken, Callum McDougall◊, Hoagy Cunningham, Thomas Henighan, Adam Jermyn, Andy Jones, Andrew Persic, Zhenyi Qi, T. Ben Thompson,

Sam Zimmerman, Kelley Rivoire, Thom…

9 months, 2 weeks назад @ youtube.com
Henry AI Labs Henry AI Labs
последний пост None
3blue1brown 3blue1brown
последний пост 1 month, 3 weeks назад
The most absurd product I've made
The most absurd product I've made The most absurd product I've made

Because why not make a pi creature neck pillow?

Available at 3b1b.co/store

1 month, 3 weeks назад @ youtube.com
How Laplace transforms solve differential equations
How Laplace transforms solve differential equations How Laplace transforms solve differential equations

Studying the forced harmonic oscillator by taking a Laplace transform and studying its poles.

Instead of sponsored ad reads, these lessons are funded directly by viewers: https://3b1b.co/support

An equally valuable form of support is to simply share the videos.

Home page: https://www.3blue1brown.com Chapter on the Laplace Transform:

https://youtu.be/j0wJBEZdwLs Chapter on the S-plane and Simple Harmonic Motion:

https://youtu.be/-j8PzkZ70Lg Timestamps:

0:00 - Opening puzzle

1:06 - Key properties of a Laplace Transform

3:29 - Qualitative analysis with Laplace Transforms

4:29 - The Laplace Transforms of a Derivative

6:06 - The forced oscillator

11:59 - Intuition from the transformed solution

1…

2 months, 1 week назад @ youtube.com
The dynamics of e^(πi)
The dynamics of e^(πi) The dynamics of e^(πi)

A fuller version of this explanation, also including the reason we care about complex exponents in the first place: https://youtu.be/-j8PzkZ70Lg

3 months назад @ youtube.com
But what is a Laplace Transform?
But what is a Laplace Transform? But what is a Laplace Transform?

Visualizing the most important tool for differential equations.

Previous chapter: https://youtu.be/-j8PzkZ70Lg

Instead of sponsored ad reads, these lessons are funded directly by viewers: https://3b1b.co/support

An equally valuable form of support is to simply share the videos.

Home page: https://www.3blue1brown.com Artwork by Kurt Bruns Engine animation borrowed with permission from this (excellent) blog: https://ciechanow.ski/internal-combustion-engine/ Timestamps:

0:00 - Understanding the engine

1:16 - Key background ideas

5:41 - Definition and intuition

10:43 - Complex integration

20:43 - Analytic continuation

23:52 - The transform of exponentials

26:15 - A deep look at cos(t)

32:59 - W…

3 months назад @ youtube.com
The dynamics of e^(πi)
The dynamics of e^(πi) The dynamics of e^(πi)

A fuller version of this explanation, also including the reason we care about complex exponents in the first place: https://youtu.be/-j8PzkZ70Lg

3 months назад @ youtube.com
Why complex exponents matter | Laplace Transform Prelude
Why complex exponents matter | Laplace Transform Prelude Why complex exponents matter | Laplace Transform Prelude

How dynamics explain Euler's formula, and vice versa.

Early view of the Laplace Transform video: https://www.patreon.com/posts/laplace-early-140428165

Instead of sponsored ad reads, these lessons are funded directly by viewers: https://3b1b.co/support

An equally valuable form of support is to simply share the videos.

Home page: https://www.3blue1brown.com Timestamps:

0:00 - Intro

1:51 - Euler's formula explained dynamically

9:27 - The harmonic oscillator

21:08 - General linear equations

22:47 - Motivating the Laplace Transform ------------------ These animations are largely made using a custom Python library, manim. See the FAQ comments here:

https://3b1b.co/faq#manim Music by Vincent Rubin…

3 months, 1 week назад @ youtube.com
Why ruler and compass? | Guest video by ⁨@bensyversen⁩
Why ruler and compass? | Guest video by ⁨@bensyversen⁩ Why ruler and compass? | Guest video by ⁨@bensyversen⁩

What role were ruler and compass constructions really serving?

Check out Ben's channel: @bensyversen Interview with the author of this video: https://youtu.be/VohYM99j8e0

Supporters get early views of new videos: https://3b1b.co/support Written, produced, edited, and animated by Ben Syversen

Additional editing: Jack Saxon

3d Blender model: Jan-Hendrik Müller

Additional Blender help: Thibaut Modrzyk (@Deepia)

Illustrations: Alex Zepherin/DonDada Studio

Drums: Jeremy Gustin

Additional music from Epidemic Sound Special thanks to Viktor Blåsjö: https://intellectualmathematics.com/opinionated-history-of-mathematics/ References/Recommended reading: Euclid’s Elements:

Visual edition of Book 1: htt…

3 months, 4 weeks назад @ youtube.com
Incomplete open cubes
Incomplete open cubes Incomplete open cubes

Full video: https://youtu.be/_BrFKp-U8GI

4 months, 1 week назад @ youtube.com
Exploration & Epiphany
Exploration & Epiphany Exploration & Epiphany

Sol Lewitt's "Incomplete Open Cubes" and rediscovering Burnside's lemma in group theory

This is a guest video by Paul Dancstep: https://youtu.be/JEeM2ABUMoo

Instead of sponsored ad reads, these lessons are funded directly by viewers: https://3b1b.co/support

An equally valuable form of support is to share the videos.

Home page: https://www.3blue1brown.com Thanks to the Wadsworth Atheneum for granting permission to use LeWitt's notebooks. Talks by Paul you can find online: What is Category Theory:

https://www.youtube.com/watch?app=desktop&v=eXBwU9ieLL0 How to Predict Eclipses:

https://www.exploratorium.edu/eclipse/video/how-predict-eclipses Theo Jansen's Strandbeests

https://www.youtube.com/w…

4 months, 1 week назад @ youtube.com
Simulating Phase Change | Guest video by Vilas Winstein
Simulating Phase Change | Guest video by Vilas Winstein Simulating Phase Change | Guest video by Vilas Winstein

Deriving the Boltzmann formula, defining temperature, and simulating liquid/vapor.

@SpectralCollective has the second part: https://youtu.be/yEcysu5xZH0

You can play with a simulation of this model here: https://vilas.us/simulations/liquidvapor/

These lessons are funded directly by viewers: https://3b1b.co/support

Home page: https://www.3blue1brown.com Notes from Vilas:

1) This open problem is to prove the ergodicity of the deterministic dynamical systems that are used to model the molecule-level physics. A good example of such a dynamical system is the box with particles evolving according to Newton's laws with elastic collisions, like in the video. 2) This video assumes that all probabili…

4 months, 2 weeks назад @ youtube.com
How AI connects text and images
How AI connects text and images How AI connects text and images

From this guest video by @WelchLabsVideo on how diffusion models work: https://youtu.be/iv-5mZ_9CPY

4 months, 3 weeks назад @ youtube.com
The AI that solved IMO Geometry Problems | Guest video by @Aleph0
The AI that solved IMO Geometry Problems | Guest video by @Aleph0 The AI that solved IMO Geometry Problems | Guest video by @Aleph0

How AlphaGeometry combines logic and intuition.

Share stories about AI in math research for an upcoming video: https://forms.gle/gr9aZVdUrW5T3yDg9

Instead of sponsored ad reads, these lessons are funded directly by viewers: https://3b1b.co/support

An equally valuable form of support is to simply share the videos.

Home page: https://www.3blue1brown.com AlphaGeometry announcement:

https://deepmind.google/discover/blog/alphageometry-an-olympiad-level-ai-system-for-geometry/ Similar open-source model, Newclid, by Harmonic:

https://harmonic.fun/news#blog-post-geometry Timestamps:

0:00 - What's surprising

1:33 - Solve without AI

7:10 - Where AI comes in

12:48 - Grant's comments ------------------…

5 months назад @ youtube.com
But how do AI videos actually work? | Guest video by @WelchLabsVideo
But how do AI videos actually work? | Guest video by @WelchLabsVideo But how do AI videos actually work? | Guest video by @WelchLabsVideo

Diffusion models, CLIP, and the math of turning text into images

Welch Labs Book: https://www.welchlabs.com/resources/imaginary-numbers-book Sections

0:00 - Intro

3:37 - CLIP

6:25 - Shared Embedding Space

8:16 - Diffusion Models & DDPM

11:44 - Learning Vector Fields

22:00 - DDIM

25:25 Dall E 2

26:37 - Conditioning

30:02 - Guidance

33:39 - Negative Prompts

34:27 - Outro

35:32 - About guest videos + Grant’s Reaction Special Thanks to:

Jonathan Ho - Jonathan is the Author of the DDPM paper and the Classifier Free Guidance Paper.

https://arxiv.org/pdf/2006.11239

https://arxiv.org/pdf/2207.12598 Preetum Nakkiran - Preetum has an excellent introductory diffusion tutorial:

https://arxiv.org/pdf/24…

5 months, 3 weeks назад @ youtube.com
Summer of Math Exposition #4 | Teachers, I'd love to hear from you
Summer of Math Exposition #4 | Teachers, I'd love to hear from you Summer of Math Exposition #4 | Teachers, I'd love to hear from you

Make a math explainer, get feedback, and receive prizes: https://some.3b1b.co

Instead of sponsored ad reads, these lessons are funded directly by viewers: https://3b1b.co/support

An equally valuable form of support is to simply share the videos. ------------------ These animations are largely made using a custom Python library, manim. See the FAQ comments here:

https://3b1b.co/faq#manim

https://github.com/3b1b/manim

https://github.com/ManimCommunity/manim/ All code for specific videos is visible here:

https://github.com/3b1b/videos/ The music is by Vincent Rubinetti.

https://www.vincentrubinetti.com

https://vincerubinetti.bandcamp.com/album/the-music-of-3blue1brown

https://open.spotify.com/…

8 months, 2 weeks назад @ youtube.com
Where my explanation of Grover’s algorithm failed
Where my explanation of Grover’s algorithm failed Where my explanation of Grover’s algorithm failed

Addressing viewer questions from the last video.

These lessons are funded directly by viewers: https://3b1b.co/support

An equally valuable form of support is to share the videos. ------------------ These animations are largely made using a custom Python library, manim. See the FAQ comments here:

https://3b1b.co/faq#manim

https://github.com/3b1b/manim

https://github.com/ManimCommunity/manim/ All code for specific videos is visible here:

https://github.com/3b1b/videos/ The music is by Vincent Rubinetti.

https://www.vincentrubinetti.com

https://vincerubinetti.bandcamp.com/album/the-music-of-3blue1brown

https://open.spotify.com/album/1dVyjwS8FBqXhRunaG5W5u ------------------ 3blue1brown is a ch…

8 months, 2 weeks назад @ youtube.com
Two Minute Papers Two Minute Papers
последний пост 2 days, 4 hours назад
Wrinkles Are Weirder Than We Ever Thought
Wrinkles Are Weirder Than We Ever Thought Wrinkles Are Weirder Than We Ever Thought

❤️ Check out Lambda here and sign up for their GPU Cloud: https://lambda.ai/papers 📝 The paper is available here:

https://wanghmin.github.io/publication/zhang-2025-pie/ Our Patreon if you wish to support us: https://www.patreon.com/TwoMinutePapers 🙏 We would like to thank our generous Patreon supporters who make Two Minute Papers possible:

Adam Bridges, Benji Rabhan, B Shang, Cameron Navor, Christian Ahlin, Eric T, Fred R, Gordon Child, Juan Benet, Michael Tedder, Owen Skarpness, Richard Sundvall, Ryan Stankye, Steef, Taras Bobrovytsky, Tazaur Sagenclaw, Tybie Fitzhugh, Ueli Gallizzi My research: https://cg.tuwien.ac.at/~zsolnai/

2 days, 4 hours назад @ youtube.com
Why Game Physics Is Falling Apart (And How To Fix It)
Why Game Physics Is Falling Apart (And How To Fix It) Why Game Physics Is Falling Apart (And How To Fix It)

❤️ Check out Lambda here and sign up for their GPU Cloud: https://lambda.ai/papers 📝 The paper is available here:

https://graphics.cs.utah.edu/research/projects/stable-cosserat-rods/ Our Patreon if you wish to support us: https://www.patreon.com/TwoMinutePapers

Note that just watching the series and leaving a kind comment every now and then is as much support as any of us could ever ask for! Sources:

https://www.youtube.com/watch?v=kO3NsSX1VTg

https://www.youtube.com/watch?v=IQZ_zBX6gQY 🙏 We would like to thank our generous Patreon supporters who make Two Minute Papers possible:

Adam Bridges, Benji Rabhan, B Shang, Cameron Navor, Christian Ahlin, Eric T, Fred R, Gordon Child, Juan Benet, Mi…

1 week назад @ youtube.com
We Just Turned Down Millions of Dollars. Here Is Why.
We Just Turned Down Millions of Dollars. Here Is Why. We Just Turned Down Millions of Dollars. Here Is Why.

Yup. My free course on how to write a light simulation program (ray tracing): https://users.cg.tuwien.ac.at/zsolnai/gfx/rendering-course/ 🙏 We would like to thank our generous Patreon supporters who make Two Minute Papers possible:

Benji Rabhan, B Shang, Christian Ahlin, Fred R, Gordon Child, Juan Benet, Michael Tedder, Owen Skarpness, Richard Sundvall, Steef, Taras Bobrovytsky, Tybie Fitzhugh, Ueli Gallizzi

If you wish to appear here or pick up other perks, click here: https://www.patreon.com/TwoMinutePapers My research: https://cg.tuwien.ac.at/~zsolnai/

Thumbnail design: Felícia Zsolnai-Fehér - http://felicia.hu

1 week, 6 days назад @ youtube.com
The Bug That Ruined Game Physics For Decades
The Bug That Ruined Game Physics For Decades The Bug That Ruined Game Physics For Decades

❤️ Check out Lambda here and sign up for their GPU Cloud: https://lambda.ai/papers Using DeepSeek on Lambda:

https://lambda.ai/inference-models/deepseek-r1 📝 The paper "A Stream Function Solver for Liquid Simulations" is available here:

https://pub.ista.ac.at/group_wojtan/projects/2015_Ando_ASFSfLS/download/vecpotential.pdf 📝 My paper on simulations that look almost like reality is available for free here:

https://rdcu.be/cWPfD Or this is the orig. Nature Physics link with clickable citations:

https://www.nature.com/articles/s41567-022-01788-5 🙏 We would like to thank our generous Patreon supporters who make Two Minute Papers possible:

Benji Rabhan, B Shang, Christian Ahlin, Fred R, Gordon …

2 weeks назад @ youtube.com
NVIDIA’s AI Learns To Walk…Painfully
NVIDIA’s AI Learns To Walk…Painfully NVIDIA’s AI Learns To Walk…Painfully

❤️ Check out Lambda here and sign up for their GPU Cloud: https://lambda.ai/papers Using DeepSeek on Lambda:

https://lambda.ai/inference-models/deepseek-r1 📝 The paper is available here:

https://research.nvidia.com/labs/toronto-ai/trace-pace/ 📝 My paper on simulations that look almost like reality is available for free here:

https://rdcu.be/cWPfD Or this is the orig. Nature Physics link with clickable citations:

https://www.nature.com/articles/s41567-022-01788-5 🙏 We would like to thank our generous Patreon supporters who make Two Minute Papers possible:

Benji Rabhan, B Shang, Christian Ahlin, Fred R, Gordon Child, Juan Benet, Michael Tedder, Owen Skarpness, Richard Sundvall, Steef, Taras B…

3 weeks, 3 days назад @ youtube.com
This Is The Physics Tech Games Have Been Waiting For
This Is The Physics Tech Games Have Been Waiting For This Is The Physics Tech Games Have Been Waiting For

❤️ Check out Weights & Biases and sign up for a free demo here: https://wandb.me/papers 📝 The paper is available here:

https://wanghmin.github.io/publication/wu-2022-gbm/ 📝 My paper on simulations that look almost like reality is available for free here:

https://rdcu.be/cWPfD Or this is the orig. Nature Physics link with clickable citations:

https://www.nature.com/articles/s41567-022-01788-5 🙏 We would like to thank our generous Patreon supporters who make Two Minute Papers possible:

Benji Rabhan, B Shang, Christian Ahlin, Fred R, Gordon Child, Juan Benet, Michael Tedder, Owen Skarpness, Richard Sundvall, Steef, Taras Bobrovytsky, Tybie Fitzhugh, Ueli Gallizzi

If you wish to appear here or …

3 weeks, 6 days назад @ youtube.com
The AI That Built An Economy… And Went Bankrupt
The AI That Built An Economy… And Went Bankrupt The AI That Built An Economy… And Went Bankrupt

❤️ Check out Lambda here and sign up for their GPU Cloud: https://lambda.ai/papers Using DeepSeek on Lambda:

https://lambda.ai/inference-models/deepseek-r1 📝 The paper is available here:

https://simworld.org/ 📝 My paper on simulations that look almost like reality is available for free here:

https://rdcu.be/cWPfD Or this is the orig. Nature Physics link with clickable citations:

https://www.nature.com/articles/s41567-022-01788-5 🙏 We would like to thank our generous Patreon supporters who make Two Minute Papers possible:

Benji Rabhan, B Shang, Christian Ahlin, Fred R, Gordon Child, Juan Benet, Michael Tedder, Owen Skarpness, Richard Sundvall, Steef, Taras Bobrovytsky, Tybie Fitzhugh, Ueli G…

1 month назад @ youtube.com
DeepMind’s Crazy New AI Masters Games That Don’t Exist
DeepMind’s Crazy New AI Masters Games That Don’t Exist DeepMind’s Crazy New AI Masters Games That Don’t Exist

❤️ Check out Lambda here and sign up for their GPU Cloud: https://lambda.ai/papers Using DeepSeek on Lambda:

https://lambda.ai/inference-models/deepseek-r1 📝 The SIMA 2 paper is available here:

https://deepmind.google/blog/sima-2-an-agent-that-plays-reasons-and-learns-with-you-in-virtual-3d-worlds/ 📝 My paper on simulations that look almost like reality is available for free here:

https://rdcu.be/cWPfD Or this is the orig. Nature Physics link with clickable citations:

https://www.nature.com/articles/s41567-022-01788-5 🙏 We would like to thank our generous Patreon supporters who make Two Minute Papers possible:

Benji Rabhan, B Shang, Christian Ahlin, Fred R, Gordon Child, Juan Benet, Michael…

1 month назад @ youtube.com
AlphaFold - The Most Important AI Breakthrough Ever Made
AlphaFold - The Most Important AI Breakthrough Ever Made AlphaFold - The Most Important AI Breakthrough Ever Made

Full interview: https://www.youtube.com/watch?v=Vhcwjzeukts

1 month назад @ youtube.com
30x Better Physics: Why Everyone Missed This Genius Solution
30x Better Physics: Why Everyone Missed This Genius Solution 30x Better Physics: Why Everyone Missed This Genius Solution

❤️ Check out Lambda here and sign up for their GPU Cloud: https://lambda.ai/papers Using DeepSeek on Lambda:

https://lambda.ai/inference-models/deepseek-r1 My hobby channel with guitars and labcoats 🥼:

https://www.youtube.com/watch?v=GjMMhn4pS38

https://www.youtube.com/watch?v=BxS62W6V48E 📝 The paper is available here:

https://arxiv.org/abs/2505.21946 📝 My paper on simulations that look almost like reality is available for free here:

https://rdcu.be/cWPfD Or this is the orig. Nature Physics link with clickable citations:

https://www.nature.com/articles/s41567-022-01788-5 🙏 We would like to thank our generous Patreon supporters who make Two Minute Papers possible:

Benji Rabhan, B Shang, Chri…

1 month, 1 week назад @ youtube.com
He Kinda Solved Biology - Nobel Prize Winner John Jumper Interview
He Kinda Solved Biology - Nobel Prize Winner John Jumper Interview He Kinda Solved Biology - Nobel Prize Winner John Jumper Interview

Thank you so much to John for being so kind and insightful, and to the film crew as well - they all did an incredible job. To celebrate the 5th anniversary of #AlphaFold, I was invited by Google DeepMind to interview Nobel Prize Winner and Distinguished Scientist, John Jumper. Note that we have no business ties with them. AlphaFold: https://deepmind.google/science/alphafold/

The full Thinking Game Movie: https://www.youtube.com/watch?v=d95J8yzvjbQ My research: https://cg.tuwien.ac.at/~zsolnai/

Thumbnail design: Felícia Zsolnai-Fehér - http://felicia.hu

1 month, 1 week назад @ youtube.com
Unreal Engine 5.7: Billions Of Triangles, In Real Time
Unreal Engine 5.7: Billions Of Triangles, In Real Time Unreal Engine 5.7: Billions Of Triangles, In Real Time

❤️ Check out Lambda here and sign up for their GPU Cloud: https://lambda.ai/papers 📝 The Unreal Engine 5.7 is available here:

https://www.unrealengine.com/en-US/news/unreal-engine-5-7-is-now-available Sources:

https://www.youtube.com/watch?v=Mj_-2SdsYLw

https://www.youtube.com/watch?v=ngzPTqtZWo4

https://advances.realtimerendering.com/s2023/2023%20Siggraph%20-%20Substrate.pdf 📝 My paper on simulations that look almost like reality is available for free here:

https://rdcu.be/cWPfD Or this is the orig. Nature Physics link with clickable citations:

https://www.nature.com/articles/s41567-022-01788-5 🙏 We would like to thank our generous Patreon supporters who make Two Minute Papers possible:

Be…

1 month, 3 weeks назад @ youtube.com
Blender 5.0 Is Here - A Revolution…For Free!
Blender 5.0 Is Here - A Revolution…For Free! Blender 5.0 Is Here - A Revolution…For Free!

❤️ Check out Lambda here and sign up for their GPU Cloud: https://lambda.ai/papers Get Blender 5.0 here: https://www.blender.org/

Example scenes: https://www.blender.org/download/demo-files/

Multiple scattering paper: https://cg.iit.bme.hu/~szirmay/volreuse_link.htm 📝 My paper on simulations that look almost like reality is available for free here:

https://rdcu.be/cWPfD Or this is the orig. Nature Physics link with clickable citations:

https://www.nature.com/articles/s41567-022-01788-5 🙏 We would like to thank our generous Patreon supporters who make Two Minute Papers possible:

Benji Rabhan, B Shang, Christian Ahlin, Gordon Child, Juan Benet, Michael Tedder, Owen Skarpness, Richard Sundvall…

1 month, 3 weeks назад @ youtube.com
DeepMind’s New AI Mastered Minecraft… Without Ever Playing It
DeepMind’s New AI Mastered Minecraft… Without Ever Playing It DeepMind’s New AI Mastered Minecraft… Without Ever Playing It

❤️ Check out Lambda here and sign up for their GPU Cloud: https://lambda.ai/papers Guide:

Rent one of their GPUs with over 16GB of VRAM

Open a terminal

Just get Ollama following the command from here - https://ollama.com/download/linux

Then run ollama run gpt-oss:120b - https://ollama.com/library/gpt-oss:120b 📝 The paper is available here:

https://danijar.com/project/dreamer4/ Source:

https://www.youtube.com/watch?v=6bnM84xGxbg 📝 My paper on simulations that look almost like reality is available for free here:

https://rdcu.be/cWPfD Or this is the orig. Nature Physics link with clickable citations:

https://www.nature.com/articles/s41567-022-01788-5 🙏 We would like to thank our generous Patre…

1 month, 3 weeks назад @ youtube.com
Games Have Never Simulated Clothing Like This Before
Games Have Never Simulated Clothing Like This Before Games Have Never Simulated Clothing Like This Before

❤️ Check out Lambda here and sign up for their GPU Cloud: https://lambda.ai/papers Guide:

Rent one of their GPUs with over 16GB of VRAM

Open a terminal

Just get Ollama with this command - https://ollama.com/download/linux

Then run ollama run gpt-oss:120b - https://ollama.com/library/gpt-oss:120b 📝 The paper "Fast Physics-Based Modeling of Knots and Ties Using Templates" is available here:

https://wanghmin.github.io/publication/guo-2025-fpb/ Sources:

https://www.youtube.com/watch?v=2RQcoLV_bVk

https://www.youtube.com/watch?v=7d158rQ1R3k

https://www.youtube.com/watch?v=qirVdKg3qgs

https://www.youtube.com/watch?v=TPokJdN2bkw

https://www.youtube.com/watch?v=DRzT3c1jk14

https://www.youtube.com/w…

1 month, 4 weeks назад @ youtube.com
DataFest Video DataFest Video
последний пост None
Семинары JetBrains Research Семинары JetBrains Research
последний пост None
Яндекс. Компьютерные науки Яндекс. Компьютерные науки
последний пост 2 weeks, 5 days назад
Компьютерное зрение в 2025-м / Роман Исаченко
Компьютерное зрение в 2025-м / Роман Исаченко Компьютерное зрение в 2025-м / Роман Исаченко

ML Global Recap 2025 — митап для ML-сообщества, на котором мы рассказали о главных международных конференциях года и самых интересных трендах в рекомендательных технологиях, компьютерном зрении, распознавании речи и NLP. С докладом на ивенте выступил Роман Исаченко, руководитель команды анализа изображений в Яндекс R&D. Он рассказал про мультимодальный анализ изображений (VLM) и диффузионки — картиночную генерацию. Больше ML-контента по ссылке: https://t.me/+Ug9D4CjJrJxmZGRi #ML, #MachineLearning, #AI, #DataScience, #MLGlobalRecap2025, #нейросети, #искусственныйинтеллект, #рекомендательныесистемы, #NLP, #ComputerVision, #SpeechRecognition, #LLM, #DeepLearning, #MLтренды, #ITконференция, #Ya…

2 weeks, 5 days назад @ youtube.com
Тренды в NLP, обзор ICLR и ACL / Александр Юшкевич
Тренды в NLP, обзор ICLR и ACL / Александр Юшкевич Тренды в NLP, обзор ICLR и ACL / Александр Юшкевич

ML Global Recap 2025 — митап для ML-сообщества, на котором мы рассказали о главных международных конференциях года и самых интересных трендах в рекомендательных технологиях, компьютерном зрении, распознавании речи и NLP. С докладом на ивенте выступил Александр Юшкевич, руководитель команды развития моделей базового качества в Поисковых сервисах и ИИ. Он показал, как конференции отражают тренды в NLP: растёт закрытость топовых LLM-моделей, а также спрос на alignment & safety, инференс, интерпретируемость и оптимизацию. А ещё появляются новые бенчмарки (куда без них). Больше ML-контента по ссылке: https://t.me/+Ug9D4CjJrJxmZGRi #ML, #MachineLearning, #AI, #DataScience, #MLGlobalRecap2025, #не…

2 weeks, 6 days назад @ youtube.com
Голосовые технологии на Interspeech и ICASSP 2025 / Борис Шелудько
Голосовые технологии на Interspeech и ICASSP 2025 / Борис Шелудько Голосовые технологии на Interspeech и ICASSP 2025 / Борис Шелудько

ML Global Recap 2025 — митап для ML-сообщества, на котором мы рассказали о главных международных конференциях года и самых интересных трендах в рекомендательных технологиях, компьютерном зрении, распознавании речи и NLP. С докладом на ивенте выступил Борис Шелудько, руководитель команды качества звука в Яндекс R&D. Он рассказал про итоги двух главных международных конференций по голосовым технологиям: Interspeech в Нидерландах и ICASSP в Индии. Больше ML-контента по ссылке: https://t.me/+Ug9D4CjJrJxmZGRi #ML, #MachineLearning, #AI, #DataScience, #MLGlobalRecap2025, #нейросети, #искусственныйинтеллект, #рекомендательныесистемы, #NLP, #ComputerVision, #SpeechRecognition, #LLM, #DeepLearning, …

3 weeks назад @ youtube.com
Главные тренды рекомендательных систем / Николай Савушкин
Главные тренды рекомендательных систем / Николай Савушкин Главные тренды рекомендательных систем / Николай Савушкин

ML Global Recap 2025 — митап для ML-сообщества, на котором мы рассказали о главных международных конференциях года и самых интересных трендах в рекомендательных технологиях, компьютерном зрении, распознавании речи и NLP. С докладом на ивенте выступил Николай Савушкин, руководитель команды рекомендательных технологий в Яндекс R&D. Он поделился инсайтами с CIKM и RecSys и рассказал про ключевые тренды в рекомендательных системах: фундаментальные и End2End-модели, масштабирование, мультимодальность, attention-based ranking и другие. Больше ML-контента по ссылке: https://t.me/+Ug9D4CjJrJxmZGRi #ML, #MachineLearning, #AI, #DataScience, #MLGlobalRecap2025, #нейросети, #искусственныйинтеллект, #ре…

3 weeks, 1 day назад @ youtube.com
Открытие ML Global Recap 2025 / Алексей Гусаков
Открытие ML Global Recap 2025 / Алексей Гусаков Открытие ML Global Recap 2025 / Алексей Гусаков

ML Global Recap 2025 — митап для ML-сообщества, на котором мы рассказали о главных международных конференциях года и самых интересных трендах в рекомендательных технологиях, компьютерном зрении, распознавании речи и NLP. Ивент открыл Алексей Гусаков, CTO Поисковых сервисов и ИИ. В докладе — вступительное слово, краткий обзор NeurIPS и немного про кризис бенчмарков. Больше ML-контента по ссылке: https://t.me/+Ug9D4CjJrJxmZGRi #ML, #MachineLearning, #AI, #DataScience, #MLGlobalRecap2025, #нейросети, #искусственныйинтеллект, #рекомендательныесистемы, #NLP, #ComputerVision, #SpeechRecognition, #LLM, #DeepLearning, #MLтренды, #ITконференция, #Yandex, #MLсообщество

3 weeks, 2 days назад @ youtube.com
Как решить проблему разнообразия ответов LLM
Как решить проблему разнообразия ответов LLM Как решить проблему разнообразия ответов LLM

Это отрывок из доклада Алексея Колесова, CTO в Яндекс R&D. На Practical ML Conf 2025 он рассказал, как ребята учили YandexGPT 5.1 лучше помнить факты и применять знания о них. А ещё показал, как у нас стабильно заработал online RL. Полная запись уже на канале! #YandexGPT #LLM #AI #ArtificialIntelligence #MachineLearning #DeepLearning #ReinforcementLearning #OnlineRL #NLP #GenerativeAI #YandexForDevelopers #YandexForML #Яндекс #AIDevDay #TechConference #DataScience #ML #AIResearch #LanguageModel #YandexTech

1 month, 1 week назад @ youtube.com
ML Global Recap'25
ML Global Recap'25 ML Global Recap'25

Митап Яндекса для ML-сообщества, на котором расскажем о шести международных конференциях и главных трендах в рекомендательных технологиях, компьютерном зрении, технологиях распознавания речи и NLP.

1 month, 1 week назад @ youtube.com
Секция на проверку базовых технических навыков ML-инженеров
Секция на проверку базовых технических навыков ML-инженеров Секция на проверку базовых технических навыков ML-инженеров

Мок-интервью: как проходит секция на базовые алгоритмы для ML-инженеров Подробнее: https://yandex.ru/jobs/interview/mldev

Вакансии: https://yandex.ru/jobs/vacancies?professions=ml-developer Подписывайтесь на телеграм-канал Яндекса для ML-сообщества: https://t.me/+Ug9D4CjJrJxmZGRi #ml, #machinelearning, #mlengineer, #mlinterview, #datascience, #яндекс, #yandex, #itcareer, #mldeveloper, #techinterview, #algorithms, #машиннообучение, #вакансии

1 month, 2 weeks назад @ youtube.com
Какие есть виды LLM-аугментаций
Какие есть виды LLM-аугментаций Какие есть виды LLM-аугментаций

Это отрывок из доклада Алексея Колесова, CTO в Яндекс R&D. На Practical ML Conf 2025 он рассказал, как ребята учили YandexGPT 5.1 лучше помнить факты и применять знания о них. А ещё показал, как у нас стабильно заработал online RL. Полная запись уже на канале! #YandexGPT #LLM #AI #ArtificialIntelligence #MachineLearning #DeepLearning #ReinforcementLearning #OnlineRL #NLP #GenerativeAI #YandexForDevelopers #YandexForML #Яндекс #AIDevDay #TechConference #DataScience #ML #AIResearch #LanguageModel #YandexTech

1 month, 2 weeks назад @ youtube.com
Как LLM предсказывает токены
Как LLM предсказывает токены Как LLM предсказывает токены

Это отрывок из доклада Алексея Колесова, CTO в Яндекс R&D. На Practical ML Conf 2025 он рассказал, как ребята учили YandexGPT 5.1 лучше помнить факты и применять знания о них. А ещё показал, как у нас стабильно заработал online RL. Полная запись уже на канале! #YandexGPT #LLM #AI #ArtificialIntelligence #MachineLearning #DeepLearning #ReinforcementLearning #OnlineRL #NLP #GenerativeAI #YandexForDevelopers #YandexForML #Яндекс #AIDevDay #TechConference #DataScience #ML #AIResearch #LanguageModel #YandexTech

1 month, 2 weeks назад @ youtube.com
«Реал Мадрид» или «Барселона»: кто круче по мнению LLM
«Реал Мадрид» или «Барселона»: кто круче по мнению LLM «Реал Мадрид» или «Барселона»: кто круче по мнению LLM

Это отрывок из доклада Алексея Колесова, CTO в Яндекс R&D. На Practical ML Conf 2025 он рассказал, как ребята учили YandexGPT 5.1 лучше помнить факты и применять знания о них. А ещё показал, как у нас стабильно заработал online RL. Полная запись уже на канале! #YandexGPT #LLM #AI #MachineLearning #GenerativeAI #YandexForML #YandexForDevelopers #Яндекс #AIDevDay #NeuralNetworks #ArtificialIntelligence #Football #Soccer #RealMadrid #Barcelona #ElClasico #LaLiga #Messi #Ronaldo #TechAndSports #AIInSports #FootballFans #SportsAnalytics #YandexTech #AIComparison

1 month, 3 weeks назад @ youtube.com
Что должен знать AI-ассистент
Что должен знать AI-ассистент Что должен знать AI-ассистент

Это отрывок из доклада Алексея Колесова, CTO в Яндекс R&D. На Practical ML Conf 2025он рассказал, как ребята учили YandexGPT 5.1 лучше помнить факты и применять знания о них. А ещё показал, как у нас стабильно заработал online RL. Полная запись уже на канале! #YandexGPT #LLM #AI #ArtificialIntelligence #MachineLearning #DeepLearning #ReinforcementLearning #OnlineRL #NLP #GenerativeAI #YandexForDevelopers #YandexForML #Яндекс #AIDevDay #TechConference #DataScience #ML #AIResearch #LanguageModel #YandexTech

1 month, 3 weeks назад @ youtube.com
Data Dojo — встреча ML-сообщества в Москве
Data Dojo — встреча ML-сообщества в Москве Data Dojo — встреча ML-сообщества в Москве

Data Dojo — это сообщество ML-экспертов. Здесь обсуждают тренды, разбирают реальные задачи, делятся опытом и практикуются. Додзё в японской культуре — место Пути, где совершенствуют не только мастерство, но и дух. Мы перенесли этот принцип в мир данных. Программа: Приветственное слово | Владислав Офицеров, модератор встречи, руководитель команды развития нейронных технологий международного поиска, и Пётр Ермаков, ML-бренд-директор Лекция: Обзор трендов и предварительный итоги года | Сергей Овчаренко, руководитель отдела мультимодального анализа и генерации Лекция: Научить AI не бредить, сдать физику и получить права: как мы готовили задачи ML‑квалификации Yandex Cup | Сергей Фиронов, ведущи…

1 month, 3 weeks назад @ youtube.com
Как обучать LLM: процесс в двух частях
Как обучать LLM: процесс в двух частях Как обучать LLM: процесс в двух частях

Это отрывок из доклада Алексея Колесова, CTO в Яндекс R&D. На Practical ML Conf 2025 он рассказал, как ребята учили YandexGPT 5.1 лучше помнить факты и применять знания о них. А ещё показал, как у нас стабильно заработал online RL. Полная запись уже на канале! #YandexGPT #LLM #AI #ArtificialIntelligence #MachineLearning #DeepLearning #ReinforcementLearning #OnlineRL #NLP #GenerativeAI #YandexForDevelopers #YandexForML #Яндекс #AIDevDay #TechConference #DataScience #ML #AIResearch #LanguageModel #YandexTech

2 months назад @ youtube.com
Визуально-языковые модели (VLM) в Яндексе: подходы, данные, подводные камни / Сергей Овчаренко
Визуально-языковые модели (VLM) в Яндексе: подходы, данные, подводные камни / Сергей Овчаренко Визуально-языковые модели (VLM) в Яндексе: подходы, данные, подводные камни / Сергей Овчаренко

Это Сергей Овчаренко, руководитель отдела мультимодальных анализа и генерации в Яндекс R&D. В своём докладе Сергей рассказал о VLM в Яндексе: какие подходы мы используем и с какими подводными камнями сталкиваемся. А еще — о претрейне и о том, почему добиться хорошего качества бывает непросто, даже когда, казалось бы, всё делаешь правильно. Узнать больше о мероприятиях для разработчиков можно тут: https://events.yandex.ru Подписывайтесь на телеграм-канал Яндекса для ML-сообщества: https://t.me/yandexforml #ML #AI #MachineLearning #DeepLearning #LLM #VLM #NeuralNetworks #Transformers #GenerativeAI #NLP #ComputerVision #DataScience #BigData #MLOps #ModelTraining #AIResearch #ArtificialIntellig…

2 months назад @ youtube.com
ML Trainings ML Trainings
последний пост 10 часов назад
Отрезвление осени и искусственный интеллект
Отрезвление осени и искусственный интеллект Отрезвление осени и искусственный интеллект 10 часов назад @ youtube.com
Нейронка и вкус: как создавать дизайн с помощью ИИ
Нейронка и вкус: как создавать дизайн с помощью ИИ Нейронка и вкус: как создавать дизайн с помощью ИИ 1 day, 10 hours назад @ youtube.com
ejbot роботы за 690 долларов из Китая
ejbot роботы за 690 долларов из Китая ejbot роботы за 690 долларов из Китая 2 days, 10 hours назад @ youtube.com
Dell отказывается от позиционирования ии
Dell отказывается от позиционирования ии Dell отказывается от позиционирования ии 3 days, 3 hours назад @ youtube.com
Капитанский мостик №1: Китайские фотоны быстрее | Роботы на детском утреннике | ПК с ИИ не нужны
Капитанский мостик №1: Китайские фотоны быстрее | Роботы на детском утреннике | ПК с ИИ не нужны Капитанский мостик №1: Китайские фотоны быстрее | Роботы на детском утреннике | ПК с ИИ не нужны

0:00:00 Начало

0:00:37 Реклама в ChatGPT

0:08:28 Manus и Цукерберг

0:14:52 Китайские фотоны быстрее

0:27:24 Новый TPU от Google

0:36:03 Рынок ускорителей от Epoch AI

0:39:52 Геймерская GPU от китайцев

0:44:09 Миллион TPU для Anthropic

0:47:39 Doubao больше всех

0:50:31 Роботы на детском утреннике

0:55:49 "Будь здоров" от ChatGPT

1:02:09 ПК с ИИ не нужны

1:10:15 Автоматизации кодинга не будет

1:18:57 Нейросети заставляют вязать ИИ-саммари: В этом выпуске обсуждаются прогнозы на новый год, реклама в ChargeGPT, покупка стартапа Manus и будущее фотонных чипов, а также новый TPU от Google. В ходе обсуждения были рассмотрены ключевые аспекты рынка чипов, включая долю рынка, хеджирование рисков и …

3 days, 10 hours назад @ youtube.com
Пузырь на рынке бихтех: мнение Дмитрия Колодезева
Пузырь на рынке бихтех: мнение Дмитрия Колодезева Пузырь на рынке бихтех: мнение Дмитрия Колодезева 1 week, 4 days назад @ youtube.com
Спутник момент: как Стивен Кинг испугался запуска
Спутник момент: как Стивен Кинг испугался запуска Спутник момент: как Стивен Кинг испугался запуска 1 week, 5 days назад @ youtube.com
Инфляция технологий и победитель получает всё
Инфляция технологий и победитель получает всё Инфляция технологий и победитель получает всё 1 week, 6 days назад @ youtube.com
Дмитрий Колодезев о реальных ограничениях и экспоненте
Дмитрий Колодезев о реальных ограничениях и экспоненте Дмитрий Колодезев о реальных ограничениях и экспоненте 2 weeks назад @ youtube.com
Дмитрий и Валентин обсуждают вайп кодинг и безопасность
Дмитрий и Валентин обсуждают вайп кодинг и безопасность Дмитрий и Валентин обсуждают вайп кодинг и безопасность 2 weeks, 1 day назад @ youtube.com
Валентин Малых говорит о проблемах адаптации технологий
Валентин Малых говорит о проблемах адаптации технологий Валентин Малых говорит о проблемах адаптации технологий 2 weeks, 2 days назад @ youtube.com
Капитанский мостик №26: Корейцы продали секрет памяти | Apple учит иcторию партии | Итоги года
Капитанский мостик №26: Корейцы продали секрет памяти | Apple учит иcторию партии | Итоги года Капитанский мостик №26: Корейцы продали секрет памяти | Apple учит иcторию партии | Итоги года

0:00:00 Начало

0:01:01 Корейцы продали секрет памяти

0:04:36 Карпаты не успевает

0:16:34 Живи быстро и не умирай

0:24:55 Бигтехи и вузы

0:35:51 Apple учит иcторию партии

0:43:32 Nvidia купила Groq

0:45:47 Япония покупает ИИ

0:53:48 Роботакси превратились в тыкву

1:00:28 Итоги года ИИ-саммари:

В этом выпуске обсуждаются ключевые события года, включая рост цен на технологии памяти, скорость изменений в индустрии, дефляцию технологий и трансформацию маркетинга. Также рассматриваются новые навыки, необходимые в программировании, и важность образования в технологической сфере России. В этом разговоре обсуждаются ключевые проблемы и перспективы агрономии, сравнение технологий в агропромышленности…

2 weeks, 3 days назад @ youtube.com
Технологии создания золотого куба
Технологии создания золотого куба Технологии создания золотого куба 2 weeks, 4 days назад @ youtube.com
Сны о дронах и киберпанке
Сны о дронах и киберпанке Сны о дронах и киберпанке 2 weeks, 5 days назад @ youtube.com
Проблемы коммуникации в большой организации
Проблемы коммуникации в большой организации Проблемы коммуникации в большой организации 2 weeks, 6 days назад @ youtube.com
Primer Primer
последний пост 2 weeks назад
Taking AI Doom Seriously For 62 Minutes
Taking AI Doom Seriously For 62 Minutes Taking AI Doom Seriously For 62 Minutes

Patreon: https://www.patreon.com/primerlearning

80,000 Hours: 80000hours.org/primer https://www.desmos.com/calculator/a5pfjtr4tr Other connections:

Discord: https://discord.gg/NbruaNW

Twitch: https://www.twitch.tv/justin_helps

Store: https://store.dftba.com/collections/primer Reddit: https://www.reddit.com/r/primerlearning/

Bsky: https://bsky.app/profile/justinhelps.bsky.social

Twitter: https://twitter.com/primerlearning Links to other resources:

https://yoshuabengio.org/2024/07/09/reasoning-through-arguments-against-taking-ai-safety-seriously/

https://www.youtube.com/c/robertmilesai

https://www.youtube.com/@Siliconversations

https://www.youtube.com/@Go-Meta

https://www.youtube.com/@Dwarkes…

2 weeks назад @ youtube.com
Simulating a single brain cell
Simulating a single brain cell Simulating a single brain cell

Patreon:

https://www.patreon.com/primerlearning Helpful resources if you want to learn more about neural networks

https://www.youtube.com/@AndrejKarpathy

https://course.fast.ai/

https://www.youtube.com/@WelchLabsVideo

https://www.youtube.com/@3blue1brown Early papers. These probably aren't helpful for understanding the concepts in this video, but if you're interested in history.

The Perceptron – A perceiving and recognizing automaton: https://bpb-us-e2.wpmucdn.com/websites.umass.edu/dist/a/27637/files/2016/03/rosenblatt-1957.pdf

The Perceptron: A probabilistic model for information storage and organization in the brain: https://www.ling.upenn.edu/courses/cogs501/Rosenblatt1958.pdf A Logical…

3 months, 2 weeks назад @ youtube.com
🎧 Podcasts
Lex Fridman AI Podcast Lex Fridman AI Podcast
последний пост 21 час назад
#489 – Paul Rosolie: Uncontacted Tribes in the Amazon Jungle
#489 – Paul Rosolie: Uncontacted Tribes in the Amazon Jungle #489 – Paul Rosolie: Uncontacted Tribes in the Amazon Jungle

Paul Rosolie is a naturalist, explorer, author of a new book titled Junglekeeper, and is someone who has dedicated his life to protecting the Amazon rainforest.

Thank you for listening ❤ Check out our sponsors: https://lexfridman.com/sponsors/ep489-scSee below for timestamps, transcript, and to give feedback, submit questions, contact Lex, etc.

Go to https://perplexity.ai/BetterHelp: Online therapy and counseling.

Go to https://fin.ai/lexMiro: Online collaborative whiteboard platform.

Go to https://miro.com/MasterClass: Online classes from world-class experts.

21 час назад @ lexfridman.com
#488 – Infinity, Paradoxes that Broke Mathematics, Gödel Incompleteness & the Multiverse – Joel David Hamkins
#488 – Infinity, Paradoxes that Broke Mathematics, Gödel Incompleteness & the Multiverse – Joel David Hamkins #488 – Infinity, Paradoxes that Broke Mathematics, Gödel Incompleteness & the Multiverse – Joel David Hamkins

Joel David Hamkins is a mathematician and philosopher specializing in set theory, the foundations of mathematics, and the nature of infinity, and he’s the #1 highest-rated user on MathOverflow.

He is also the author of several books, including Proof and the Art of Mathematics and Lectures on the Philosophy of Mathematics.

Thank you for listening ❤ Check out our sponsors: https://lexfridman.com/sponsors/ep488-scSee below for timestamps, transcript, and to give feedback, submit questions, contact Lex, etc.

Go to https://masterclass.com/lexpodOUTLINE:(00:00) – Introduction(01:58) – Sponsors, Comments, and Reflections(15:40) – Infinity & paradoxes(1:02:50) – Russell’s paradox(1:15:57) – Gödel’s…

1 week, 6 days назад @ lexfridman.com
#487 – Irving Finkel: Deciphering Secrets of Ancient Civilizations & Flood Myths
#487 – Irving Finkel: Deciphering Secrets of Ancient Civilizations & Flood Myths #487 – Irving Finkel: Deciphering Secrets of Ancient Civilizations & Flood Myths

Irving Finkel is a scholar of ancient languages and a longtime curator at the British Museum, renowned for his expertise in Mesopotamian history and cuneiform writing.

He specializes in reading and interpreting cuneiform inscriptions, including tablets from Sumerian, Akkadian, Babylonian, and Assyrian contexts.

Thank you for listening ❤ Check out our sponsors: https://lexfridman.com/sponsors/ep487-scSee below for timestamps, transcript, and to give feedback, submit questions, contact Lex, etc.

Go to https://shopify.com/lexMiro: Online collaborative whiteboard platform.

Go to https://miro.com/Chevron: Reliable energy for data centers.

1 month назад @ lexfridman.com
#486 – Michael Levin: Hidden Reality of Alien Intelligence & Biological Life
#486 – Michael Levin: Hidden Reality of Alien Intelligence & Biological Life #486 – Michael Levin: Hidden Reality of Alien Intelligence & Biological Life

Michael Levin is a biologist at Tufts University working on novel ways to understand and control complex pattern formation in biological systems.

Thank you for listening ❤ Check out our sponsors: https://lexfridman.com/sponsors/ep486-scSee below for timestamps, transcript, and to give feedback, submit questions, contact Lex, etc.

Go to https://upliftdesk.com/lexMiro: Online collaborative whiteboard platform.

Go to https://miro.com/MasterClass: Online classes from world-class experts.

(2:42:41) – Mind uploading(3:01:22) – Alien intelligence(3:16:17) – Advice for young people(3:22:46) – Questions for AGI

1 month, 2 weeks назад @ lexfridman.com
#485 – David Kirtley: Nuclear Fusion, Plasma Physics, and the Future of Energy
#485 – David Kirtley: Nuclear Fusion, Plasma Physics, and the Future of Energy #485 – David Kirtley: Nuclear Fusion, Plasma Physics, and the Future of Energy

David Kirtley is a nuclear fusion engineer and CEO of Helion Energy, a company working on building the world's first commercial fusion power plant by 2028.

Thank you for listening ❤ Check out our sponsors: https://lexfridman.com/sponsors/ep485-sc

See below for timestamps, transcript, and to give feedback, submit questions, contact Lex, etc. Transcript:

https://lexfridman.com/david-kirtley-transcript CONTACT LEX:

Feedback - give feedback to Lex: https://lexfridman.com/survey

AMA - submit questions, videos or call-in: https://lexfridman.com/ama

Hiring - join our team: https://lexfridman.com/hiring

Other - other ways to get in touch: https://lexfridman.com/contact EPISODE LINKS:

David's X: htt…

1 month, 3 weeks назад @ lexfridman.com
#484 – Dan Houser: GTA, Red Dead Redemption, Rockstar, Absurd & Future of Gaming
#484 – Dan Houser: GTA, Red Dead Redemption, Rockstar, Absurd & Future of Gaming #484 – Dan Houser: GTA, Red Dead Redemption, Rockstar, Absurd & Future of Gaming

Dan Houser is co-founder of Rockstar Games and is a legendary creative mind behind Grand Theft Auto (GTA) and Red Dead Redemption series of video games.

Thank you for listening ❤ Check out our sponsors: https://lexfridman.com/sponsors/ep484-scSee below for timestamps, transcript, and to give feedback, submit questions, contact Lex, etc.

Go to https://box.com/aiUPLIFT Desk: Standing desks and office ergonomics.

Go to https://drinkLMNT.com/lexOUTLINE:(00:00) – Introduction(01:29) – Sponsors, Comments, and Reflections(11:32) – Greatest films of all time(23:45) – Making video games(26:36) – GTA 3(29:55) – Open world video games(32:42) – Character creation(36:09) – Superintelligent AI in A Bette…

2 months, 2 weeks назад @ lexfridman.com
#483 – Julia Shaw: Criminal Psychology of Murder, Serial Killers, Memory & Sex
#483 – Julia Shaw: Criminal Psychology of Murder, Serial Killers, Memory & Sex #483 – Julia Shaw: Criminal Psychology of Murder, Serial Killers, Memory & Sex

Julia Shaw is a criminal psychologist and author who in her books explores human nature, including psychopathy, violent crime, the psychology of evil, police interrogation, false memory manipulation, deception detection, and human sexuality.

Thank you for listening ❤ Check out our sponsors: https://lexfridman.com/sponsors/ep483-scSee below for timestamps, transcript, and to give feedback, submit questions, contact Lex, etc.

Go to https://shopify.com/lexBetterHelp: Online therapy and counseling.

Go to https://betterhelp.com/lexLMNT: Zero-sugar electrolyte drink mix.

Go to https://drinkLMNT.com/lexAG1: All-in-one daily nutrition drink.

3 months назад @ lexfridman.com
#482 – Pavel Durov: Telegram, Freedom, Censorship, Money, Power & Human Nature
#482 – Pavel Durov: Telegram, Freedom, Censorship, Money, Power & Human Nature #482 – Pavel Durov: Telegram, Freedom, Censorship, Money, Power & Human Nature

Pavel Durov is the founder and CEO of Telegram.

Thank you for listening ❤ Check out our sponsors: https://lexfridman.com/sponsors/ep482-scSee below for timestamps, transcript, and to give feedback, submit questions, contact Lex, etc.

Transcript:https://lexfridman.com/pavel-durov-transcriptCONTACT LEX:Feedback – give feedback to Lex: https://lexfridman.com/surveyAMA – submit questions, videos or call-in: https://lexfridman.com/amaHiring – join our team: https://lexfridman.com/hiringOther – other ways to get in touch: https://lexfridman.com/contactEPISODE LINKS:Pavel’s Telegram: https://t.me/durovPavel’s X: https://x.com/durovTelegram: https://telegram.org/Telegram Contests: https://contest.c…

3 months, 2 weeks назад @ lexfridman.com
#481 – Norman Ohler: Hitler, Nazis, Drugs, WW2, Blitzkrieg, LSD, MKUltra & CIA
#481 – Norman Ohler: Hitler, Nazis, Drugs, WW2, Blitzkrieg, LSD, MKUltra & CIA #481 – Norman Ohler: Hitler, Nazis, Drugs, WW2, Blitzkrieg, LSD, MKUltra & CIA

Norman Ohler is a historian and author of “Blitzed: Drugs in the Third Reich,” a book that investigates the role of psychoactive drugs, particularly stimulants such as methamphetamine, in the military history of World War II.

It is a book that two legendary historians Ian Kershaw and Antony Beevor give very high praise for its depth of research.

Norman also wrote “Tripped: Nazi Germany, the CIA, and the Dawn of the Psychedelic Age”, and he is working on a new book “Stoned Sapiens” looking at the history of human civilization through the lens of drugs.

Thank you for listening ❤ Check out our sponsors: https://lexfridman.com/sponsors/ep481-scSee below for timestamps, transcript, and to give f…

3 months, 3 weeks назад @ lexfridman.com
#480 – Dave Hone: T-Rex, Dinosaurs, Extinction, Evolution, and Jurassic Park
#480 – Dave Hone: T-Rex, Dinosaurs, Extinction, Evolution, and Jurassic Park #480 – Dave Hone: T-Rex, Dinosaurs, Extinction, Evolution, and Jurassic Park

Dave Hone is a paleontologist, expert on dinosaurs, co-host of the Terrible Lizards podcast, and author of numerous scientific papers and books on the behavior and ecology of dinosaurs.

He lectures at Queen Mary University of London on topics of Ecology, Zoology, Biology, and Evolution.

Thank you for listening ❤ Check out our sponsors: https://lexfridman.com/sponsors/ep480-scSee below for timestamps, transcript, and to give feedback, submit questions, contact Lex, etc.

Go to https://go.lindy.ai/lexBetterHelp: Online therapy and counseling.

Go to https://shopify.com/lexLMNT: Zero-sugar electrolyte drink mix.

4 months, 1 week назад @ lexfridman.com
#479 – Dave Plummer: Programming, Autism, and Old-School Microsoft Stories
#479 – Dave Plummer: Programming, Autism, and Old-School Microsoft Stories #479 – Dave Plummer: Programming, Autism, and Old-School Microsoft Stories

Dave Plummer is a programmer, former Microsoft software engineer (Windows 95, NT, XP), creator of Task Manager, author of two books on autism, and host of the Dave’s Garage YouTube channel, where he shares stories from his career, insights on software development, and deep dives into technology.

Thank you for listening ❤ Check out our sponsors: https://lexfridman.com/sponsors/ep479-scSee below for timestamps, and to give feedback, submit questions, contact Lex, etc.

Go to https://upliftdesk.com/lexZocDoc: App that helps patients find healthcare providers.

Go to https://zocdoc.com/lexFin: AI agent for customer service.

Go to https://fin.ai/lexAllio Capital: AI-powered investment app that use…

4 months, 2 weeks назад @ lexfridman.com
#478 – Scott Horton: The Case Against War and the Military Industrial Complex
#478 – Scott Horton: The Case Against War and the Military Industrial Complex #478 – Scott Horton: The Case Against War and the Military Industrial Complex

Scott Horton is the director of the Libertarian Institute, editorial director of Antiwar.com, host of The Scott Horton Show, co-host of Provoked, and for the past three decades a staunch critic of U.S. military interventionism.

Thank you for listening ❤ Check out our sponsors: https://lexfridman.com/sponsors/ep478-scSee below for timestamps, and to give feedback, submit questions, contact Lex, etc.

Go to https://alliocapital.com/Hampton: Community for high-growth founders and CEOs.

Go to https://joinhampton.com/lexBetterHelp: Online therapy and counseling.

Go to https://drinkag1.com/lexOUTLINE:(00:00) – Introduction(00:35) – Sponsors, Comments, and Reflections(09:14) – From the Cold War to …

4 months, 3 weeks назад @ lexfridman.com
#477 – Keyu Jin: China’s Economy, Tariffs, Trade, Trump, Communism & Capitalism
#477 – Keyu Jin: China’s Economy, Tariffs, Trade, Trump, Communism & Capitalism #477 – Keyu Jin: China’s Economy, Tariffs, Trade, Trump, Communism & Capitalism

Keyu Jin is an economist specializing in China’s economy, international macroeconomics, global trade imbalances, and financial policy.

She is the author of The New China Playbook: Beyond Socialism and Capitalism.

Thank you for listening ❤ Check out our sponsors: https://lexfridman.com/sponsors/ep477-scSee below for timestamps, transcript, and to give feedback, submit questions, contact Lex, etc.

Go to https://alliocapital.com/UPLIFT Desk: Standing desks and office ergonomics.

Go to https://upliftdesk.com/lexHampton: Community for high-growth founders and CEOs.

5 months назад @ lexfridman.com
#476 – Jack Weatherford: Genghis Khan and the Mongol Empire
#476 – Jack Weatherford: Genghis Khan and the Mongol Empire #476 – Jack Weatherford: Genghis Khan and the Mongol Empire

Jack Weatherford is an anthropologist and historian specializing in Genghis Khan and the Mongol Empire.

Thank you for listening ❤ Check out our sponsors: https://lexfridman.com/sponsors/ep476-scSee below for timestamps, and to give feedback, submit questions, contact Lex, etc.

Go to https://alliocapital.com/ZocDoc: App that helps patients find healthcare providers.

Go to https://zocdoc.com/lexFin: AI agent for customer service.

Go to https://shopify.com/lexMasterClass: Online classes from world-class experts.

5 months, 2 weeks назад @ lexfridman.com
#475 – Demis Hassabis: Future of AI, Simulating Reality, Physics and Video Games
#475 – Demis Hassabis: Future of AI, Simulating Reality, Physics and Video Games #475 – Demis Hassabis: Future of AI, Simulating Reality, Physics and Video Games

Demis Hassabis is the CEO of Google DeepMind and Nobel Prize winner for his groundbreaking work in protein structure prediction using AI.

Thank you for listening ❤ Check out our sponsors: https://lexfridman.com/sponsors/ep475-scSee below for timestamps, transcript, and to give feedback, submit questions, contact Lex, etc.

Go to https://joinhampton.com/lexFin: AI agent for customer service.

Go to https://shopify.com/lexLMNT: Zero-sugar electrolyte drink mix.

Go to https://drinkLMNT.com/lexAG1: All-in-one daily nutrition drink.

5 months, 3 weeks назад @ lexfridman.com
Microsoft Research Podcast Microsoft Research Podcast
последний пост 1 month, 1 week назад
Ideas: Community building, machine learning, and the future of AI
Ideas: Community building, machine learning, and the future of AI Ideas: Community building, machine learning, and the future of AI

This week, machine learning researchers around the world will be attending the annual Conference on Neural Information Processing Systems, or NeurIPS.

In this series, we’ll explore the technologies that are shaping our future and the big ideas that propel them forward.

So around that time when I started my PhD at Penn, I was working in machine learning theory and algorithmic economics.

How had you experienced a lack of community or network of women in machine learning before the founding of WiML?

So particularly when working on topics related to fairness, I’ve ended up focusing a bunch on stuff to do with marginalized groups as part of my responsible AI work.

1 month, 1 week назад @ microsoft.com
Ideas: More AI-resilient biosecurity with the Paraphrase Project
Ideas: More AI-resilient biosecurity with the Paraphrase Project Ideas: More AI-resilient biosecurity with the Paraphrase Project

Today, I’m excited to talk about the Paraphrase Project, an effort I co-led exploring how advances in AI tools for protein design might impact biosecurity.

These “patches,” akin to those in cybersecurity, have now been shared with organizations globally to strengthen biosecurity screening.

The project highlights that the same AI tools capable of incredible good can also be misused, requiring us to be vigilant, thoughtful, and creative so we continue to get the most benefit out of AI tools while working to ensure that we avoid costly misuses.

So things like, how similar is this to that template, wild-type protein structure that we used as our conditioning information?

But I feel like broadly…

3 months, 1 week назад @ microsoft.com
Coauthor roundtable: Reflecting on healthcare economics, biomedical research, and medical education
Coauthor roundtable: Reflecting on healthcare economics, biomedical research, and medical education Coauthor roundtable: Reflecting on healthcare economics, biomedical research, and medical education

KOHANE: So I think you’ve “nerd sniped” me because you [LAUGHTER]—which is all too easy—but I think there’s a central issue here.

But I actually think this is dark matter of human organizational technology that is not well understood.

AZEEM AZHAR: We didn’t talk about, you know, AI in its ability to potentially do this, which is to extend the clinician’s presence throughout the week.

And so I think there’s always going to be an opening for either differences of opinion or agreeing with you too much.

And this gets into whether AI is really going to get almost to the ab initio understanding of human biology.

4 months, 3 weeks назад @ microsoft.com
Reimagining healthcare delivery and public health with AI
Reimagining healthcare delivery and public health with AI Reimagining healthcare delivery and public health with AI

We are sorry, the page you requested cannot be found.

The page you are looking for could not be found or is no longer available.

5 months, 1 week назад @ microsoft.com
Navigating medical education in the era of generative AI
Navigating medical education in the era of generative AI Navigating medical education in the era of generative AI

Prior to med school, Daniel pursued experiences that cultivated his interest in the application of AI in medical practice and education.

Really, really looking forward to this chat.

There’s AI before ChatGPT and before, you know, generative AI really became a big thing, and then afterwards.

And then after we talk about what’s really happening, what do you think should happen in medical education given the reality of generative AI?

And I do agree [that] AI really gives us real hope that we can make it true.

5 months, 3 weeks назад @ microsoft.com
AI Testing and Evaluation: Reflections
AI Testing and Evaluation: Reflections AI Testing and Evaluation: Reflections

Our goal is to learn from their successes and their stumbles to move the science and practice of AI testing forward.

We have examples, like the pharmaceutical or medical device industry experts with whom you spoke, that’s really, you know, testing … there is a pre-deployment requirement.

And the third is just how rigid versus adaptive these testing and evaluation regimes or frameworks are in these different domains.

I really agree that there has been a lot of emphasis to date on, sort of, testing models upstream, the AI model evaluation.

You know, I think there’s been real progress already in the AI evaluation and testing ecosystem in the public-private partnership context.

5 months, 3 weeks назад @ microsoft.com
AI Testing and Evaluation: Learnings from cybersecurity
AI Testing and Evaluation: Learnings from cybersecurity AI Testing and Evaluation: Learnings from cybersecurity

Absolutely, I really, really was.

As a principal director on the Microsoft AI Red Team, Tori leads all AI security and safety red team operations, as well as dangerous capability testing, to directly inform C-suite decision-makers.

This year, we’ve pulled a lot of those assets and insights into the Azure [AI] Foundry AI Red Teaming Agent (opens in new tab).

So you can get a little taste of what we do day to day in the AI Red Teaming Agent.

WESTERHOFF: I think the most important takeaway from those lessons is that AI security is truly a team sport.

6 months назад @ microsoft.com
How AI will accelerate biomedical research and discovery
How AI will accelerate biomedical research and discovery How AI will accelerate biomedical research and discovery

Dr. Eric Topol is the executive vice president of the biomedical research non-profit Scripps Research, where he founded and now directs the Scripps Research Translational Institute.

Let’s continue our deep dive on AI and biomedical research with this conversation with Noubar Afeyan:LEE: Noubar, thanks so much for joining.

And there’s the origin story of contact with AI, you know, before the emergence of generative AI and afterwards.

What is going on today with respect to AI really being used for something meaningful in the design and development of drugs?

TOPOL: You would read about how, you know, data is the new oil and, you know, gold and whatnot.

6 months, 1 week назад @ microsoft.com
AI Testing and Evaluation: Learnings from pharmaceuticals and medical devices
AI Testing and Evaluation: Learnings from pharmaceuticals and medical devices AI Testing and Evaluation: Learnings from pharmaceuticals and medical devices

Our goal is to learn from their successes and their stumbles to move the science and practice of AI testing forward.

During the pre-market phase, medical testing establishes baseline safety and effectiveness metrics through bench testing, performance standards, and clinical studies.

SULLIVAN: So medical devices face a pretty prescriptive multi-level testing path before they hit the market.

We are looking into medical devices, as well, obviously, but also other technologies in advanced medical computing.

So we see Phase 3 trials as something that occurs in the medical devices and pharmaceuticals field.

6 months, 1 week назад @ microsoft.com
AI Testing and Evaluation: Learnings from genome editing
AI Testing and Evaluation: Learnings from genome editing AI Testing and Evaluation: Learnings from genome editing

As generative AI continues to advance, Microsoft has gathered a range of experts—from genome editing to cybersecurity—to share how their fields approach evaluation and risk assessment.

CHARO: Well, you know, genome editing is both very old and very new.

Now the earliest forms of genome editing were very inefficient, and so we didn’t worry that much.

But the bottom-line thing to remember, the way to really think about it is, we don’t regulate genome editing; we regulate the things that use genome editing.

And she said, you know, we don’t regulate genome editing; we regulate the things that use genome editing.

6 months, 2 weeks назад @ microsoft.com
AI Testing and Evaluation: Learnings from Science and Industry
AI Testing and Evaluation: Learnings from Science and Industry AI Testing and Evaluation: Learnings from Science and Industry

Our goal is to learn from their successes and their stumbles to move the science and practice of AI testing forward.

And I think, really, there are two reasons why tech is so, kind of, representative of that kind of challenge that I’ve always found fascinating.

Continues to be a really important topic in the AI policy conversation right now, I think, for really good reason.

Testing is an important component for governance and AI and, of course, in all of these other domains, as well.

I think about almost, like, in the near to mid-term, like three issues that we need to address in the AI, kind of, policy and testing context.

6 months, 3 weeks назад @ microsoft.com
The AI Revolution in Medicine, Revisited: How AI is reshaping the future of healthcare and medical research
The AI Revolution in Medicine, Revisited: How AI is reshaping the future of healthcare and medical research The AI Revolution in Medicine, Revisited: How AI is reshaping the future of healthcare and medical research

LEE: Yeah, yeah.

It cannot—as, you know, Bill was saying—it cannot learn from your document.

And I don’t know if the two of you remember, but I ended up doing a lot of tests.

I don’t know if you know, but just recently, there was a paper that was published on a scientific discovery using o3- mini (opens in new tab).

Like, if you have a human trained for one task and you put them into another task, then you don’t … you often don’t know.

7 months назад @ microsoft.com
What AI's impact on individuals means for the health workforce and industry
What AI's impact on individuals means for the health workforce and industry What AI's impact on individuals means for the health workforce and industry

So I don’t think we should be surprised that business schools matter on this because we care about management.

That’s really going to change the way, like, middle school works, was my thinking at the time.

We’ve gone from AI being highly discriminative to AI that’s able to explore the world in particular ways.

The symptoms that they’re showing are quite different, and also their compliance is really, really different.

LEE: Yeah, really, really interesting.

7 months, 2 weeks назад @ microsoft.com
Abstracts: Zero-shot models in single-cell biology with Alex Lu
Abstracts: Zero-shot models in single-cell biology with Alex Lu Abstracts: Zero-shot models in single-cell biology with Alex Lu

And single-cell foundation models claim to be capable of unraveling deeper insights than ever before.

Basically, we showed that single-cell foundation models perform worse in settings that are fundamental to biological discovery than much simpler machine learning and statistical methods that were used in the field before single-cell foundation models emerged and are the go-to standard for unpacking meaning from these complicated experiments.

And the way to understand this is because single-cell foundation models are trained in a way that tries to expose these models to millions of single-cells.

But let’s also talk about the impact for methodologists, people who are trying to improve these s…

7 months, 3 weeks назад @ microsoft.com
Abstracts: Aurora with Megan Stanley and Wessel Bruinsma
Abstracts: Aurora with Megan Stanley and Wessel Bruinsma Abstracts: Aurora with Megan Stanley and Wessel Bruinsma

This is such exciting work about environmental forecasting, so we’re happy to have the two of you join us today.

Mostly because AI weather forecasting models are computationally much more efficient and can even be more accurate.

What’s unfortunate though, about this big step forward, is that these developments are mostly limited to the setting of weather forecasting.

Weather forecasting is very important, obviously, but there are many other important environmental forecasting problems out there, such as air pollution forecasting or ocean wave forecasting.

STANLEY: Current approaches have really focused training very specifically on weather forecasting models.

7 months, 4 weeks назад @ microsoft.com
NLP Highlights NLP Highlights
последний пост None
Data Skeptic
последний пост 2 weeks, 5 days назад
Video Recommendations in Industry
Video Recommendations in Industry Video Recommendations in Industry

In this episode, Kyle Polich sits down with Cory Zechmann, a content curator working in streaming television with 16 years of experience running the music blog "Silence Nogood." They explore the intersection of human curation and machine learning in content discovery, discussing the concept of "algatorial" curation—where algorithms and editorial expertise work together. Key topics include the cold start problem, why every metric is just a "proxy metric" for what users actually want, the challenge of filter bubbles, and the importance of balancing familiarity with discovery. Cory shares insights on why TikTok's algorithm works so well (clean data and massive interaction volume), the crucial …

2 weeks, 5 days назад @ dataskeptic.com
Eye Tracking in Recommender Systems
Eye Tracking in Recommender Systems Eye Tracking in Recommender Systems

In this episode, Santiago de Leon takes us deep into the world of eye tracking and its revolutionary applications in recommender systems. As a researcher at the Kempelin Institute and Brno University, Santiago explains the mechanics of eye tracking technology—how it captures gaze data and processes it into fixations and saccades to reveal user browsing patterns. He introduces the groundbreaking RecGaze dataset, the first eye tracking dataset specifically designed for recommender systems research, which opens new possibilities for understanding how users interact with carousel interfaces like Netflix. Through collaboration between psychologists and AI researchers, Santiago's work demonstrate…

3 weeks, 6 days назад @ dataskeptic.com
Cracking the Cold Start Problem
Cracking the Cold Start Problem Cracking the Cold Start Problem

In this episode of Data Skeptic, we dive deep into the technical foundations of building modern recommender systems. Unlike traditional machine learning classification problems where you can simply apply XGBoost to tabular data, recommender systems require sophisticated hybrid approaches that combine multiple techniques. Our guest, Boya Xu, an assistant professor of marketing at Virginia Tech, walks us through a cutting-edge method that integrates three key components: collaborative filtering for dimensionality reduction, embeddings to represent users and items in latent space, and bandit learning to balance exploration and exploitation when deploying new recommendations. Boya shares insigh…

1 month, 1 week назад @ dataskeptic.com
Designing Recommender Systems for Digital Humanities
Designing Recommender Systems for Digital Humanities Designing Recommender Systems for Digital Humanities

In this episode of Data Skeptic, we explore the fascinating intersection of recommender systems and digital humanities with guest Florian Atzenhofer-Baumgartner, a PhD student at Graz University of Technology. Florian is working on Monasterium.net, Europe's largest online collection of historical charters, containing millions of medieval and early modern documents from across the continent. The conversation delves into why traditional recommender systems fall short in the digital humanities space, where users range from expert historians and genealogists to art historians and linguists, each with unique research needs and information-seeking behaviors. Florian explains the technical challen…

1 month, 3 weeks назад @ dataskeptic.com
DataRec Library for Reproducible in Recommend Systems
DataRec Library for Reproducible in Recommend Systems DataRec Library for Reproducible in Recommend Systems

In this episode of Data Skeptic's Recommender Systems series, host Kyle Polich explores DataRec, a new Python library designed to bring reproducibility and standardization to recommender systems research. Guest Alberto Carlo Mario Mancino, a postdoc researcher from Politecnico di Bari, Italy, discusses the challenges of dataset management in recommendation research—from version control issues to preprocessing inconsistencies—and how DataRec provides automated downloads, checksum verification, and standardized filtering strategies for popular datasets like MovieLens, Last.fm, and Amazon reviews. The conversation covers Alberto's research journey through knowledge graphs, graph-based recommen…

2 months назад @ dataskeptic.com
Shilling Attacks on Recommender Systems
Shilling Attacks on Recommender Systems Shilling Attacks on Recommender Systems

In this episode of Data Skeptic's Recommender Systems series, Kyle sits down with Aditya Chichani, a senior machine learning engineer at Walmart, to explore the darker side of recommendation algorithms. The conversation centers on shilling attacks—a form of manipulation where malicious actors create multiple fake profiles to game recommender systems, either to promote specific items or sabotage competitors. Aditya, who researched these attacks during his undergraduate studies at SPIT before completing his master's in computer science with a data science specialization at UC Berkeley, explains how these vulnerabilities emerge particularly in collaborative filtering systems. From promoting a …

2 months, 1 week назад @ dataskeptic.com
Music Playlist Recommendations
Music Playlist Recommendations Music Playlist Recommendations

In this episode, Rebecca Salganik, a PhD student at the University of Rochester with a background in vocal performance and composition, discusses her research on fairness in music recommendation systems. She explores three key types of fairness—group, individual, and counterfactual—and examines how algorithms create challenges like popularity bias (favoring mainstream content) and multi-interest bias (underserving users with diverse tastes). Rebecca introduces LARP, her multi-stage multimodal framework for playlist continuation that uses contrastive learning to align text and audio representations, learn song relationships, and create playlist-level embeddings to address the cold start prob…

2 months, 2 weeks назад @ dataskeptic.com
Bypassing the Popularity Bias
Bypassing the Popularity Bias Bypassing the Popularity Bias 3 months назад @ dataskeptic.com
Sustainable Recommender Systems for Tourism
Sustainable Recommender Systems for Tourism Sustainable Recommender Systems for Tourism

In this episode, we speak with Ashmi Banerjee, a doctoral candidate at the Technical University of Munich, about her pioneering research on AI-powered recommender systems in tourism. Ashmi illuminates how these systems can address exposure bias while promoting more sustainable tourism practices through innovative approaches to data acquisition and algorithm design. Key highlights include leveraging large language models for synthetic data generation, developing recommendation architectures that balance user satisfaction with environmental concerns, and creating frameworks that distribute tourism more equitably across destinations. Ashmi's insights offer valuable perspectives for both AI res…

3 months, 1 week назад @ dataskeptic.com
Interpretable Real Estate Recommendations
Interpretable Real Estate Recommendations Interpretable Real Estate Recommendations

In this episode of Data Skeptic's Recommender Systems series, host Kyle Polich interviews Dr. Kunal Mukherjee, a postdoctoral research associate at Virginia Tech, about the paper "Z-REx: Human-Interpretable GNN Explanations for Real Estate Recommendations" The discussion explores how the post-COVID real estate landscape has created a need for better recommendation systems that can introduce home buyers to emerging neighborhoods they might not know about. Dr. Mukherjee, explains how his team developed a graph neural network approach that not only recommends properties but provides human-interpretable explanations for why certain regions are suggested. The conversation covers the advantages o…

3 months, 3 weeks назад @ dataskeptic.com
Why Am I Seeing This?
Why Am I Seeing This? Why Am I Seeing This?

In this episode of Data Skeptic, we explore the challenges of studying social media recommender systems when exposure data isn't accessible. Our guests Sabrina Guidotti, Gregor Donabauer, and Dimitri Ognibene introduce their innovative "recommender neutral user model" for inferring the influence of opaque algorithms.

4 months, 1 week назад @ dataskeptic.com
Eco-aware GNN Recommenders
Eco-aware GNN Recommenders Eco-aware GNN Recommenders

In this episode of Data Skeptic, we dive into eco-friendly AI with Antonio Purificato, a PhD student from Sapienza University of Rome. Antonio discusses his research on "EcoAware Graph Neural Networks for Sustainable Recommendations" and explores how we can measure and reduce the environmental impact of recommender systems without sacrificing performance.

4 months, 2 weeks назад @ dataskeptic.com
Networks and Recommender Systems
Networks and Recommender Systems Networks and Recommender Systems

Kyle reveals the next season's topic will be "Recommender Systems". Asaf shares insights on how network science contributes to the recommender system field.

4 months, 4 weeks назад @ dataskeptic.com
Network of Past Guests Collaborations
Network of Past Guests Collaborations Network of Past Guests Collaborations

Kyle and Asaf discuss a project in which we link former guests of the podcast based on their co-authorship of academic papers.

5 months, 3 weeks назад @ dataskeptic.com
The Network Diversion Problem
The Network Diversion Problem The Network Diversion Problem

In this episode, Professor Pål Grønås Drange from the University of Bergen, introduces the field of Parameterized Complexity - a powerful framework for tackling hard computational problems by focusing on specific structural aspects of the input. This framework allows researchers to solve NP-complete problems more efficiently when certain parameters, like the structure of the graph, are "well-behaved". At the center of the discussion is the network diversion problem, where the goal isn’t to block all routes between two points in a network, but to force flow - such as traffic, electricity, or data - through a specific path. While this problem appears deceptively similar to the classic "Min.Cu…

6 months, 1 week назад @ dataskeptic.com
SuperDataScience SuperDataScience
последний пост 1 day, 5 hours назад
957: How AI Agents Are Automating Enterprise Data Operations, with Ashwin Rajeeva
957: How AI Agents Are Automating Enterprise Data Operations, with Ashwin Rajeeva 957: How AI Agents Are Automating Enterprise Data Operations, with Ashwin Rajeeva

AI agents, data lakes, and managing data sprawl: Ashwin Rajeeva, cofounder and CTO of Acceldata, speaks to Jon Krohn about how the agentic data management startup raised over $100 million in venture capital to expand its business in automating data quality assurance as well as cataloguing and pipeline maintenance across enterprise environments. Acceldata utilizes multiple agents to solve enterprise-grade questions with company data. It also uses autonomous data pipelines that can detect and fix issues without human intervention, and the platform’s agentic data management system ADM also lets humans stay in the loop wherever needed. This episode is brought to you by the ⁠⁠Dell⁠⁠, by ⁠⁠Intel⁠…

1 day, 5 hours назад @ podtrac.com
956: From Agent Demo to Enterprise Product (with Ease!) feat. Salesforce’s Tyler Carlson
956: From Agent Demo to Enterprise Product (with Ease!) feat. Salesforce’s Tyler Carlson 956: From Agent Demo to Enterprise Product (with Ease!) feat. Salesforce’s Tyler Carlson

#Sponsored SVP, Head of Product for Salesforce’s AppExchange & Ecosystem, Tyler Carlson, talks to Jon Krohn about taking AI agents from prototype to enterprise-grade production with the Agentforce 360 Platform. Though we may now have plenty of tools to build demos for AI agents, most teams still struggle to turn early prototypes into secure and scalable products. With Salesforce’s Agentforce 360 Platform, users can build customer-focused agentic applications with a multi-LLM planner service for reasoning and logic, as well as a new scripting language for deterministic control over how agents interact with the contextual layer of the Salesforce data model. Learn how Salesforce’s WYSIWYG sche…

5 days, 5 hours назад @ podtrac.com
955: Nested Learning, Spatial Intelligence and the AI Trends of 2026, with Sadie St. Lawrence
955: Nested Learning, Spatial Intelligence and the AI Trends of 2026, with Sadie St. Lawrence 955: Nested Learning, Spatial Intelligence and the AI Trends of 2026, with Sadie St. Lawrence

Sadie St Lawrence joins Jon Krohn to discuss what to expect from the AI industry in 2026. Sadie and Jon talk through what they think will be the five biggest trends in AI, hand out awards for the best moments, comebacks, and disappointments in AI in 2025, and review how their predictions for 2025 played out. Hear Sadie’s five exciting predictions for 2026, from emerging jobs in AI to an important return to the drawing board! This episode is brought to you by the ⁠⁠Dell⁠⁠, by ⁠⁠Intel⁠⁠, by Fabi and by MongoDB. Additional materials: ⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠www.superdatascience.com/955⁠⁠⁠⁠⁠ Interested in sponsoring a SuperDataScience Podcast episode? Email [email protected] for spon…

1 week, 1 day назад @ podtrac.com
954: Recap of 2025 and Wishing You a Wonderful 2026
954: Recap of 2025 and Wishing You a Wonderful 2026 954: Recap of 2025 and Wishing You a Wonderful 2026

Jon Krohn wraps up 2025 with his thoughts on how agentic AI has become as much a resounding success as an annoying buzzword for many in the tech industry, why such promising developments in generative AI mean that well-prepared, secured data will be ever more crucial, and Jon’s hopes for a better year for everyone across the world in 2026. Additional materials: ⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠www.superdatascience.com/954⁠⁠⁠⁠⁠⁠⁠⁠ Interested in sponsoring a SuperDataScience Podcast episode? Email [email protected] for sponsorship information.

1 week, 5 days назад @ podtrac.com
953: Beyond “Agent Washing”: AI Systems That Actually Deliver ROI, with Dell’s Global CTO John Roese
953: Beyond “Agent Washing”: AI Systems That Actually Deliver ROI, with Dell’s Global CTO John Roese 953: Beyond “Agent Washing”: AI Systems That Actually Deliver ROI, with Dell’s Global CTO John Roese

Dell Technologies’ John Roese talks to Jon Krohn about the phenomenon of “agent-washing”, his contribution to Dell’s incredible revenue boost in 2025, and why “knowledge layers” will be crucial to future tech. Hear also John’s predictions for where AI is going to lead us in 2026, from better, clearer governance, data management methods and definitions for agentic AI, to systems that keep AI tools and our data running and secure with the help of “AI factories” and “sovereign AI”. This episode is brought to you by MongoDB and by Y Carrot. Additional materials: ⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠www.superdatascience.com/953⁠⁠⁠⁠⁠⁠ Interested in sponsoring a SuperDataScience Podcast episode? Email natalie…

2 weeks, 1 day назад @ podtrac.com
952: How to Avoid Burnout and Get Promoted, with “The Fit Data Scientist” Penelope Lafeuille
952: How to Avoid Burnout and Get Promoted, with “The Fit Data Scientist” Penelope Lafeuille 952: How to Avoid Burnout and Get Promoted, with “The Fit Data Scientist” Penelope Lafeuille

“The Fit Data Scientist” newsletter author Pénélope Lafeuille talks to Jon Krohn about how to give your all at work, offering her top tips for a healthy body and a healthy mind. Learn why “The SuperDataScience Podcast” made it onto her top 3 data science podcasts, and why following your passion can pay off in dividends for your career. Additional materials: ⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠www.superdatascience.com/952⁠⁠⁠⁠⁠⁠ Interested in sponsoring a SuperDataScience Podcast episode? Email [email protected] for sponsorship information.

2 weeks, 5 days назад @ podtrac.com
951: Context Engineering, Multiplayer AI and Effective Search, with Dropbox’s Josh Clemm
951: Context Engineering, Multiplayer AI and Effective Search, with Dropbox’s Josh Clemm 951: Context Engineering, Multiplayer AI and Effective Search, with Dropbox’s Josh Clemm

VP of Engineering at Dropbox Josh Clemm speaks to Jon Krohn about consolidating search tools across apps with the AI-powered workspace, Dropbox Dash, the new collaborative AI systems that enhance interoperability between team members and their projects, and how to avoid “context rot”. Dropbox Dash gives users the best of Dropbox’s cloud storage and search functions, plus a “universal search” ability to locate information across multimedia and apps. “AI really needs to understand you and your team, first and foremost, and all that connected data,” says Josh. This episode is brought to you by the ⁠⁠Dell⁠⁠, by ⁠⁠Intel⁠⁠, by Airia and by MongoDB. Additional materials: ⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠www…

3 weeks, 1 day назад @ podtrac.com
950: Happy Holidays from All of Us at the SuperDataScience Podcast
950: Happy Holidays from All of Us at the SuperDataScience Podcast 950: Happy Holidays from All of Us at the SuperDataScience Podcast

In this special holiday episode, the SuperDataScience Podcast team comes together to wish you happy holidays and thank you for listening throughout the year. Team members from around the world share warm greetings in their own voices and languages as we reflect on another year of learning, curiosity, and community. From all of us at SDS, we wish you a joyful holiday season and look forward to bringing you more data science, machine learning, and AI content in the year ahead. Additional materials: ⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠www.superdatascience.com/950⁠⁠⁠⁠⁠⁠⁠ Interested in sponsoring a SuperDataScience Podcast episode? Email [email protected] for sponsorship information.

3 weeks, 5 days назад @ podtrac.com
949: Why AI Keeps Failing Society, with Stanford professor Alex “Sandy” Pentland
949: Why AI Keeps Failing Society, with Stanford professor Alex “Sandy” Pentland 949: Why AI Keeps Failing Society, with Stanford professor Alex “Sandy” Pentland

Alex “Sandy” Pentland, Toshiba Professor of Media Arts & Science at MIT and Fellow at Stanford, speaks to Jon Krohn about his new book, Shared Wisdom, why he attributes AI to the collapse of the Soviet Union, and why those risks to society could still be relevant today. We can only achieve better system performance, Alex says, when we build tools that keep step with the way that people make decisions. Listen to the episode to hear Alex talk about how he is helping make AI agents work for individuals rather than the companies that develop them, and his work in making sure that systems operate consistently and fairly across the world. This episode is brought to you by the⁠ ⁠⁠⁠⁠Dell⁠⁠⁠, by⁠ ⁠⁠…

4 weeks, 1 day назад @ podtrac.com
948: In Case You Missed It in November 2025
948: In Case You Missed It in November 2025 948: In Case You Missed It in November 2025

In this November episode of “In Case You Missed It” series, Jon Krohn selects his favorite clips from the month. Hear from Shirish Gupta and Tyler Cox (Episode 939), Vikoy Pandey (Episode 941), Marc Dupuis (Episode 937), and Maya Ackerman (Episode 943) on getting back to human motivation and the importance of evaluating the tools and data we use. Additional materials: ⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠www.superdatascience.com/948⁠⁠⁠⁠⁠⁠⁠⁠⁠ Interested in sponsoring a SuperDataScience Podcast episode? Email [email protected] for sponsorship information.

1 month назад @ podtrac.com
947: How to Get Hired at Top Firms like Netflix and Spotify, with Jeff Li
947: How to Get Hired at Top Firms like Netflix and Spotify, with Jeff Li 947: How to Get Hired at Top Firms like Netflix and Spotify, with Jeff Li

Jeff Li tells Jon Krohn what it's like to work at scale as a data scientist and a machine learning engineer at Netflix, Spotify and DoorDash, as well as how to get a foot in the door at these companies. Jeff also discusses how to run forecasts and trends, and how to read their results. Listen to hear Jeff Li discuss how Spotify became a podcast powerhouse, his startup move.ai, and the tools he uses every day. This episode is brought to you by the ⁠⁠Dell⁠⁠, by ⁠⁠Intel⁠⁠, by Fabi, and by Airia. Additional materials: ⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠www.superdatascience.com/947⁠⁠⁠⁠⁠ Interested in sponsoring a SuperDataScience Podcast episode? Email [email protected] for sponsorship informat…

1 month назад @ podtrac.com
946: How Robotaxis Are Transforming Cities
946: How Robotaxis Are Transforming Cities 946: How Robotaxis Are Transforming Cities

Jon Krohn looks into the benefits of robotaxis, from safety to affordability, in this Five-Minute Friday. Hear about Waymo’s partnership with Jaguar Land Rover, the latest safety studies concerning driverless vehicles, and a case for robotaxis becoming the preferred method of transport in the US, where households spend roughly 15% of their budget on vehicle ownership. Additional materials: ⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠www.superdatascience.com/946⁠⁠⁠⁠⁠⁠ Interested in sponsoring a SuperDataScience Podcast episode? Email [email protected] for sponsorship information.

1 month, 1 week назад @ podtrac.com
945: AI is a Joke, with Joel Beasley
945: AI is a Joke, with Joel Beasley 945: AI is a Joke, with Joel Beasley

Is there humor in data? Joel Beasley, host of Modern CTO, tells Jon Krohn how he used AI to turn his sights to stand-up comedy. He also shares his tips on tech leadership that he learned from his popular podcast, Modern CTO, and how he is using generative AI as a collaborative partner in his creative work. This episode is brought to you by the ⁠⁠Dell⁠⁠, by ⁠⁠Intel⁠⁠, by Fabi, and by Gurobi⁠⁠⁠. Additional materials: ⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠www.superdatascience.com/945⁠⁠⁠⁠ Interested in sponsoring a SuperDataScience Podcast episode? Email [email protected] for sponsorship information. In this episode you will learn: (02:14) Joel Beasley on his comedy career (19:04) Applying the ‘me…

1 month, 1 week назад @ podtrac.com
944: Gemini 3 Pro: Google’s Back on Top
944: Gemini 3 Pro: Google’s Back on Top 944: Gemini 3 Pro: Google’s Back on Top

Google is steaming ahead with launching its top-league new Gemini 3 Pro model across their product suite, from Google Search to Vertex AI cloud services. The multinational tech company is also letting eager early adopters like Wayfair and GitHub. Get all the detailed data, its performance across hard-to-game industry benchmarks, and what this all means for the way you use generative AI, in this week’s episode. Additional materials: ⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠www.superdatascience.com/944⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠ Interested in sponsoring a SuperDataScience Podcast episode? Email [email protected] for sponsorship information.

1 month, 2 weeks назад @ podtrac.com
943: Creative Machines: AI in Music and Art, with Prof. Maya Ackerman
943: Creative Machines: AI in Music and Art, with Prof. Maya Ackerman 943: Creative Machines: AI in Music and Art, with Prof. Maya Ackerman

Creative human-AI partnerships and AI-generated music: WaveAI CEO and co-founder Maya Ackerman speaks with Jon Krohn about learning to see – and accept – AI’s potential as a creative partner in a human-centric, AI-forward future. Listen to the episode to hear Maya Ackerman discuss reframing hallucination as a creative force, her work at WaveAI, and how to push the boundaries of creativity using generative AI. This episode is brought to you by the ⁠⁠Dell⁠⁠, by ⁠⁠Intel⁠⁠, by Gurobi⁠⁠⁠ and by Airia. Additional materials: ⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠www.superdatascience.com/943⁠⁠⁠ Interested in sponsoring a SuperDataScience Podcast episode? Email [email protected] for sponsorship informat…

1 month, 2 weeks назад @ podtrac.com
Data Science at Home Data Science at Home
последний пост 3 weeks, 1 day назад
When Data Stops Being Code and Starts Being Conversation (Ep. 297)
When Data Stops Being Code and Starts Being Conversation (Ep. 297) When Data Stops Being Code and Starts Being Conversation (Ep. 297)

Mark Brocato built Mockaroo—the tool that taught millions of developers how to fake data.

Now, as Head of Engineering at Tonic.ai, he’s building the AI agent that’s making his own creation obsolete.

From the hidden failures of legacy mocks to the security implications of agent-driven synthesis, Mark reveals what happens when data generation becomes a conversation—not a pipeline.

SponsorsTonic.ai Synthetic data solutions for software and AI development.

Accelerate engineering velocity and ensure compliance with AI-powered data synthesisThis episode is brought to you by Statistical HorizonsAt Statistical Horizons, you can stay ahead with expert-led livestream seminars that make data analytics…

3 weeks, 1 day назад @ datascienceathome.com
Your AI Strategy is Burning Money: Here’s How to Fix It (Ep.295)
Your AI Strategy is Burning Money: Here’s How to Fix It (Ep.295) Your AI Strategy is Burning Money: Here’s How to Fix It (Ep.295)

Most companies don’t have an AI problem.

In this conversation, he breaks down when AI actually makes sense, where AWS costs spiral out of control, and why your “cool demo” keeps dying before launch.

If you’re tired of AI hype and ready for straight answers, hit play.

Our Discord community is full of ML engineers, researchers, and AI enthusiasts discussing papers, sharing projects, and helping each other level up.

Whether you’re debugging your first neural net or training your tenth transformer, there’s a place for you.

1 month, 2 weeks назад @ datascienceathome.com
From Tokens to Vectors: The Efficiency Hack That Could Save AI (Ep. 294)
From Tokens to Vectors: The Efficiency Hack That Could Save AI (Ep. 294) From Tokens to Vectors: The Efficiency Hack That Could Save AI (Ep. 294)

LLMs generate text painfully slow, one low-info token at a time.

Researchers just figured out how to compress 4 tokens into smart vectors & cut costs by 44%—with full code & proofs!

🔥📊SponsorsThis episode is brought to you by Statistical HorizonsAt Statistical Horizons, you can stay ahead with expert-led livestream seminars that make data analytics and AI methods practical and accessible.

Join thousands of researchers and professionals who’ve advanced their careers with Statistical Horizons.

Get $200 off any seminar with code DATA25 at https://statisticalhorizons.com

2 months назад @ datascienceathome.com
Why AI Researchers Are Suddenly Obsessed With Whirlpools (Ep. 293)
Why AI Researchers Are Suddenly Obsessed With Whirlpools (Ep. 293) Why AI Researchers Are Suddenly Obsessed With Whirlpools (Ep. 293)

VortexNet uses actual whirlpools to build neural networks.

By borrowing equations from fluid dynamics, this new architecture might solve deep learning’s toughest problems—from vanishing gradients to long-range dependencies.

Today we explain how vortex shedding, the Strouhal number, and turbulent flows might change everything in AI.

SponsorsThis episode is brought to you by Statistical HorizonsAt Statistical Horizons, you can stay ahead with expert-led livestream seminars that make data analytics and AI methods practical and accessible.

Join thousands of researchers and professionals who’ve advanced their careers with Statistical Horizons.

2 months, 2 weeks назад @ datascienceathome.com
The Scientists Growing Living Computers in Swiss Labs (Ep. 292)
The Scientists Growing Living Computers in Swiss Labs (Ep. 292) The Scientists Growing Living Computers in Swiss Labs (Ep. 292)

At the intersection of ethics and engineering, Amethix creates AI systems that don’t just function—they adapt, learn, and serve.

With a focus on dual-use innovation, Amethix is shaping a future where intelligent machines extend human capability, not replace it.

Discover more at https://amethix.com This episode is brought to you by Intrepid AI.

From drones to satellites, Intrepid AI gives engineers and defense innovators the tools to prototype, simulate, and deploy autonomous systems with confidence.

Learn more at intrepid.aiReferencesWebsite: finalspark.comDiscord account: / discordNewsletter: https://finalspark.com/#newsletterTopics: Biological computing • Neural engineering • Energy-effic…

2 months, 3 weeks назад @ datascienceathome.com
When AI Hears Thunder But Misses the Fear (Ep. 291)
When AI Hears Thunder But Misses the Fear (Ep. 291) When AI Hears Thunder But Misses the Fear (Ep. 291)

Sanjoy Chowdhury reveals AI’s hidden weakness: while systems can see objects and hear sounds perfectly, they can’t reason across senses like humans do.

At the intersection of ethics and engineering, Amethix creates AI systems that don’t just function—they adapt, learn, and serve.

Discover more at https://amethix.comThis episode is brought to you by Intrepid AI.

From drones to satellites, Intrepid AI gives engineers and defense innovators the tools to prototype, simulate, and deploy autonomous systems with confidence.

Whether it’s in the sky, on the ground, or in orbit—if it’s intelligent and mobile, Intrepid helps you build it.

3 months, 1 week назад @ datascienceathome.com
Why VCs Are Funding $100M Remote Control Toys (Ep. 290)
Why VCs Are Funding $100M Remote Control Toys (Ep. 290) Why VCs Are Funding $100M Remote Control Toys (Ep. 290)

ReferencesWar On The Rocks: https://warontherocks.com/2025/08/ukraine-isnt-the-model-for-winning-the-innovation-war/LinkedIn: https://www.linkedin.com/in/jonasrsinger/Spotify: https://tr.ee/Omy_1X8k1UApple Podcast: https://podcasts.apple.com/us/podcast/defence-innovation-podcast/id1797131332YouTube: https://youtube.com/@DefenceInnovationpodcast?si=cu2WlnVgL5XKnM0pSponsorsThis episode is proudly sponsored by Amethix Technologies.

At the intersection of ethics and engineering, Amethix creates AI systems that don’t just function—they adapt, learn, and serve.

Discover more at https://amethix.comThis episode is brought to you by Intrepid AI.

From drones to satellites, Intrepid AI gives engineers…

3 months, 4 weeks назад @ datascienceathome.com
How Hacker Culture Died (Ep. 289)
How Hacker Culture Died (Ep. 289) How Hacker Culture Died (Ep. 289)

At the intersection of ethics and engineering, Amethix creates AI systems that don’t just function—they adapt, learn, and serve.

Discover more at amethix.comDSH is brought to you by Intrepid AI.

🐦 Twitter: @DataScienceAtHome📘 LinkedIn: https://www.linkedin.com/in/fragadaleta/Instagram: https://www.instagram.com/datascienceathome/Facebook: https://www.facebook.com/datascienceAHLinkedIn: https://www.linkedin.com/company/data-science-at-home-podcastDiscord Channel: https://discord.gg/4UNKGf3NEW TO DATA SCIENCE AT HOME?

Data Science at Home explores the latest in AI, data science, and machine learning.

Send us mail at:[email protected]’t forget to like, subscribe, and hit the 🔔 for…

4 months, 2 weeks назад @ datascienceathome.com
Robots Suck (But It’s Not Their Fault) (Ep. 288)
Robots Suck (But It’s Not Their Fault) (Ep. 288) Robots Suck (But It’s Not Their Fault) (Ep. 288)

At the intersection of ethics and engineering, Amethix creates AI systems that don’t just function—they adapt, learn, and serve.

Discover more at amethix.comDSH is brought to you by Intrepid AI.

🐦 Twitter: @DataScienceAtHome📘 LinkedIn: https://www.linkedin.com/in/fragadaleta/Instagram: https://www.instagram.com/datascienceathome/Facebook: https://www.facebook.com/datascienceAHLinkedIn: https://www.linkedin.com/company/data-science-at-home-podcastDiscord Channel: https://discord.gg/4UNKGf3NEW TO DATA SCIENCE AT HOME?

Data Science at Home explores the latest in AI, data science, and machine learning.

Send us mail at:[email protected]’t forget to like, subscribe, and hit the 🔔 for…

5 months, 1 week назад @ datascienceathome.com
Your Favorite AI Startup is Probably Bullshit (Ep. 287)
Your Favorite AI Startup is Probably Bullshit (Ep. 287) Your Favorite AI Startup is Probably Bullshit (Ep. 287)

The brutal truth about why Silicon Valley is blowing billions on glorified autocomplete while pretending it’s the next iPhone.

We’re diving deep into the AI investment circus where VCs who can’t code are funding companies that barely understand their own technology.

From blockchain déjà vu to the “ChatGPT wrapper” economy—this episode will make you question every AI valuation you’ve ever seen.

Fair warning: We’re naming names and calling out the hype.

Don’t listen if you work at a “revolutionary AI startup” that’s just OpenAI’s API with a pretty interface.

5 months, 2 weeks назад @ datascienceathome.com
Tech’s Dumbest Mistake: Why Firing Programmers for AI Will Destroy Everything (Ep. 286) [RB]
Tech’s Dumbest Mistake: Why Firing Programmers for AI Will Destroy Everything (Ep. 286) [RB] Tech’s Dumbest Mistake: Why Firing Programmers for AI Will Destroy Everything (Ep. 286) [RB]

From the viral article “Tech’s Dumbest Mistake: Why Firing Programmers for AI Will Destroy Everything” on my newsletter at https://defragzone.substack.com/p/techs-dumbest-mistake-why-firinghere are my thoughts about AI replacing programmers…🎙️ Sponsors AGNTCY — The open source collective building the Internet of Agents🌐 https://www.agntcy.org✨ Connect with us!

🐦 Twitter: @DataScienceAtHome📘 LinkedIn: https://www.linkedin.com/in/fragadaleta/Instagram: https://www.instagram.com/datascienceathome/Facebook: https://www.facebook.com/datascienceAHLinkedIn: https://www.linkedin.com/company/data-science-at-home-podcastDiscord Channel: https://discord.gg/4UNKGf3NEW TO DATA SCIENCE AT HOME?

Data Scie…

6 months, 1 week назад @ datascienceathome.com
Brains in the Machine: The Rise of Neuromorphic Computing (Ep. 285)
Brains in the Machine: The Rise of Neuromorphic Computing (Ep. 285) Brains in the Machine: The Rise of Neuromorphic Computing (Ep. 285)

In this episode of Data Science at Home, we explore the fascinating world of neuromorphic computing — a brain-inspired approach to computation that could reshape the future of AI and robotics.

The episode breaks down how neuromorphic systems differ from conventional AI architectures like transformers and LLMs, diving into spiking neural networks (SNNs), their benefits in energy efficiency and real-time processing, and their limitations in training and scalability.

Real-world applications are highlighted, including low-power drones, hearing aids, and event-based cameras.

Francesco closes with a vision of hybrid systems where neuromorphic chips and LLMs coexist, blending biological inspiratio…

7 months назад @ datascienceathome.com
DSH/Warcoded – AI in the Invisible Battlespace (Ep. 284)
DSH/Warcoded – AI in the Invisible Battlespace (Ep. 284) DSH/Warcoded – AI in the Invisible Battlespace (Ep. 284)

This episode explores the invisible battlespace of cyber and electronic warfare, where AI takes center stage.

SponsorsBuilding multi-agent software is hard — agent-to-agent and agent-to-tool communication is still the wild west.

At the intersection of ethics and engineering, Amethix creates AI systems that don’t just function—they adapt, learn, and serve.

Discover more at amethix.comWarcoded is brought to you by Intrepid AI.

From drones to satellites, Intrepid AI gives engineers and defense innovators the tools to prototype, simulate, and deploy autonomous systems with confidence.

7 months, 2 weeks назад @ datascienceathome.com
DSH/Warcoded Swarming the Battlefield (Ep. 283)
DSH/Warcoded Swarming the Battlefield (Ep. 283) DSH/Warcoded Swarming the Battlefield (Ep. 283)

Swarming the Battlefield explores how artificial intelligence is revolutionizing combat through coordinated drone swarms.

This episode uncovers how these intelligent agents turn the chaos of the battlefield into a synchronized dance of machine warfare.

At the intersection of ethics and engineering, Amethix creates AI systems that don’t just function—they adapt, learn, and serve.

Discover more at amethix.comWarcoded is brought to you by Intrepid AI.

From drones to satellites, Intrepid AI gives engineers and defense innovators the tools to prototype, simulate, and deploy autonomous systems with confidence.

7 months, 3 weeks назад @ datascienceathome.com
DSH/Warcoded Kill Chains and Algorithmic Warfare – Autonomy in Targeting and Engagement (Ep. 282)
DSH/Warcoded Kill Chains and Algorithmic Warfare – Autonomy in Targeting and Engagement (Ep. 282) DSH/Warcoded Kill Chains and Algorithmic Warfare – Autonomy in Targeting and Engagement (Ep. 282)

In this gripping follow-up, we dive into how AI is transforming kinetic operations—from identifying a threat to executing a strike.

At the intersection of ethics and engineering, Amethix creates AI systems that don’t just function—they adapt, learn, and serve.

Discover more at amethix.comWarcoded is brought to you by Intrepid AI.

From drones to satellites, Intrepid AI gives engineers and defense innovators the tools to prototype, simulate, and deploy autonomous systems with confidence.

Whether it’s in the sky, on the ground, or in orbit—if it’s intelligent and mobile, Intrepid helps you build it.

8 months назад @ datascienceathome.com