Discussion
Loading...

#Tag

  • About
  • Code of conduct
  • Privacy
  • Users
  • Instances
  • About Bonfire
phildini
@phildini@wandering.shop  ·  activity timestamp 2 months ago

Inspired by a talk I had with @BajoranEngineer at #PyTorchCon, I've jotted down some thoughts about #Python as a scripting engine for apps.

https://phildini.dev/python-in-every-app

Shares appreciated! Commentary welcome, but if you're a jerk I'll block you 😇

@freakboy3742 @glyph @brettcannon this is why I was asking about built python ✨

Also included: a thought on how @conda monetizes this 😅

  • Copy link
  • Flag this post
  • Block
phildini
@phildini@wandering.shop  ·  activity timestamp 2 months ago

Moving from Training to Inference, we have Peng Wu from Meta.

The trends here match the trends in training! Heterogeneous hardware, dsitributed inference, concern over numerics and determinism -- all on the rise!

Also seeing deep consolidation in LLM serving platforms.

#PyTorchCon

phildini
@phildini@wandering.shop replied  ·  activity timestamp 2 months ago

PyTorch Compiler is a control point for all of these. Compiler can apply LLM-specific and CUDA-specific optimizations to code.

#PyTorchCon

  • Copy link
  • Flag this comment
  • Block
phildini
@phildini@wandering.shop  ·  activity timestamp 2 months ago

The PyTorch team is also trying to solve specific problems! Like: numerics-sensitive models that make compilers sussy.

The solution here is region-based inductors with in-code annotations.

#PyTorchCon

phildini
@phildini@wandering.shop replied  ·  activity timestamp 2 months ago

Moving from Training to Inference, we have Peng Wu from Meta.

The trends here match the trends in training! Heterogeneous hardware, dsitributed inference, concern over numerics and determinism -- all on the rise!

Also seeing deep consolidation in LLM serving platforms.

#PyTorchCon

  • Copy link
  • Flag this comment
  • Block
phildini
@phildini@wandering.shop  ·  activity timestamp 2 months ago

What does the modern stack look like for post-training RL?

Training, Inference, and Rollouts all happen under one roof, all with their own orchestration needs.

One solution to all this: Monarch (from Meta?)

#PyTorchCon

phildini
@phildini@wandering.shop replied  ·  activity timestamp 2 months ago

The PyTorch team is also trying to solve specific problems! Like: numerics-sensitive models that make compilers sussy.

The solution here is region-based inductors with in-code annotations.

#PyTorchCon

  • Copy link
  • Flag this comment
  • Block
phildini
@phildini@wandering.shop  ·  activity timestamp 2 months ago

It's been a wild year! DeepSeek R1 was this past January, proving out the value of mixture of experts, reinforcement learning, and the shift from pre-training to post-training inference.

We also saw a huge rise in coding agents, multimodal deployment, and AI at the edge.

#PyTorchCon

phildini
@phildini@wandering.shop replied  ·  activity timestamp 2 months ago

What does the modern stack look like for post-training RL?

Training, Inference, and Rollouts all happen under one roof, all with their own orchestration needs.

One solution to all this: Monarch (from Meta?)

#PyTorchCon

  • Copy link
  • Flag this comment
  • Block
phildini
@phildini@wandering.shop  ·  activity timestamp 2 months ago

Now: trends in PyTorch for Training with Edward Yang from Meta.

#PyTorchCon

phildini
@phildini@wandering.shop replied  ·  activity timestamp 2 months ago

It's been a wild year! DeepSeek R1 was this past January, proving out the value of mixture of experts, reinforcement learning, and the shift from pre-training to post-training inference.

We also saw a huge rise in coding agents, multimodal deployment, and AI at the edge.

#PyTorchCon

  • Copy link
  • Flag this comment
  • Block
phildini
@phildini@wandering.shop  ·  activity timestamp 2 months ago

What's coolest is that there's over 1000 contributors to the codebase over the past 12 months!

The follow on effect is that 400k projects were added to github over the past twelve months that use PyTorch

#PyTorchCon

phildini
@phildini@wandering.shop replied  ·  activity timestamp 2 months ago

Now: trends in PyTorch for Training with Edward Yang from Meta.

#PyTorchCon

  • Copy link
  • Flag this comment
  • Block
phildini
@phildini@wandering.shop  ·  activity timestamp 2 months ago

PyTorch has a dominant position with AI researchers. 16k papers published in the past year using PyTorch.

Community involvement also remains high! 90k stars / 25k forks on github. This leads to 14k commits made to PyTorch core over the past year

#PyTorchCon

phildini
@phildini@wandering.shop replied  ·  activity timestamp 2 months ago

What's coolest is that there's over 1000 contributors to the codebase over the past 12 months!

The follow on effect is that 400k projects were added to github over the past twelve months that use PyTorch

#PyTorchCon

  • Copy link
  • Flag this comment
  • Block
phildini
@phildini@wandering.shop  ·  activity timestamp 2 months ago

Next Keynote speaker! Alban Desmaison (and others!) for a #PyTorch Technical Deep Dive.

This is the PyTorch Core Maintainers talk ✨

#PyTorchCon

phildini
@phildini@wandering.shop replied  ·  activity timestamp 2 months ago

PyTorch has a dominant position with AI researchers. 16k papers published in the past year using PyTorch.

Community involvement also remains high! 90k stars / 25k forks on github. This leads to 14k commits made to PyTorch core over the past year

#PyTorchCon

  • Copy link
  • Flag this comment
  • Block
phildini
@phildini@wandering.shop  ·  activity timestamp 2 months ago

Next Keynote speaker! Alban Desmaison (and others!) for a #PyTorch Technical Deep Dive.

This is the PyTorch Core Maintainers talk ✨

#PyTorchCon

  • Copy link
  • Flag this post
  • Block
phildini
@phildini@wandering.shop  ·  activity timestamp 2 months ago

Next speaker: Anush Elangovan from AMD talking about Empowering Developers with AMD ROCm™ Everywhere.

#PyTorchCon

phildini
@phildini@wandering.shop replied  ·  activity timestamp 2 months ago

ROCm is a full-stack approach for open development in machine learning and LLMs.

Three pillars:
- Ci/CD for streamlined dev
- OSS projects for first-class support
- one-click deploy of jupyter noteboooks to AMD GPUs

#PyTorchCon

  • Copy link
  • Flag this comment
  • Block
phildini
@phildini@wandering.shop  ·  activity timestamp 2 months ago

Next speaker: Anush Elangovan from AMD talking about Empowering Developers with AMD ROCm™ Everywhere.

#PyTorchCon

  • Copy link
  • Flag this post
  • Block
phildini
@phildini@wandering.shop  ·  activity timestamp 2 months ago

Wrapping up, we need more hardware-efficient AI training.

"Hardware should be part of the constraint of algorithm design"

The true bottleneck right now is memory. Memory is a bigger bottleneck than compute.

#PyTorchCon

phildini
@phildini@wandering.shop replied  ·  activity timestamp 2 months ago

Lots of free lunches still exist for efficiency! Being more hardware-aware will produce better results soon.

#PyTorchCon

  • Copy link
  • Flag this comment
  • Block
phildini
@phildini@wandering.shop  ·  activity timestamp 2 months ago

Neural networks need to be able to understand mathematical reasoning -- LEAN / LeanDojo are critical steps in this path.

This team took LEAN and built an LLM out of it, they're hoping to Open Source soon

#PyTorchCon

phildini
@phildini@wandering.shop replied  ·  activity timestamp 2 months ago

Wrapping up, we need more hardware-efficient AI training.

"Hardware should be part of the constraint of algorithm design"

The true bottleneck right now is memory. Memory is a bigger bottleneck than compute.

#PyTorchCon

  • Copy link
  • Flag this comment
  • Block
phildini
@phildini@wandering.shop  ·  activity timestamp 2 months ago

Beyond modeling phenomena, these neural operators are able to do hardware designed at the quantum gate level.

(I absolutely do not know enough about the science here, but it looks cool!)

#PyTorchCon

phildini
@phildini@wandering.shop replied  ·  activity timestamp 2 months ago

Neural networks need to be able to understand mathematical reasoning -- LEAN / LeanDojo are critical steps in this path.

This team took LEAN and built an LLM out of it, they're hoping to Open Source soon

#PyTorchCon

  • Copy link
  • Flag this comment
  • Block
phildini
@phildini@wandering.shop  ·  activity timestamp 2 months ago

They've completed a weather forecasting model that is 45,000 times faster than current models (not sure how measured).

The model is Apache-licensed #PyTorch code.

#PyTorchCon

phildini
@phildini@wandering.shop replied  ·  activity timestamp 2 months ago

Beyond modeling phenomena, these neural operators are able to do hardware designed at the quantum gate level.

(I absolutely do not know enough about the science here, but it looks cool!)

#PyTorchCon

  • Copy link
  • Flag this comment
  • Block
phildini
@phildini@wandering.shop  ·  activity timestamp 2 months ago

NeuralOperator is the group this work is based on, check them out on github. There will also be a poster at #PyTorchCon

Now on to success stories.

phildini
@phildini@wandering.shop replied  ·  activity timestamp 2 months ago

They've completed a weather forecasting model that is 45,000 times faster than current models (not sure how measured).

The model is Apache-licensed #PyTorch code.

#PyTorchCon

  • Copy link
  • Flag this comment
  • Block
phildini
@phildini@wandering.shop  ·  activity timestamp 2 months ago

The team has developed Physics-Informed Neural Operators (PINO).

Can more accurately predict phenomena like fluid dynamics (very cool demo that's hard to capture in text)

#PyTorchCon

phildini
@phildini@wandering.shop replied  ·  activity timestamp 2 months ago

NeuralOperator is the group this work is based on, check them out on github. There will also be a poster at #PyTorchCon

Now on to success stories.

  • Copy link
  • Flag this comment
  • Block
phildini
@phildini@wandering.shop  ·  activity timestamp 2 months ago

We need new neural operators that can handle physical, mathematical domains better.

Using lower-res translations and models considered harmful! Think: Hurricanes

#PyTorchCon TorchCon

phildini
@phildini@wandering.shop replied  ·  activity timestamp 2 months ago

The team has developed Physics-Informed Neural Operators (PINO).

Can more accurately predict phenomena like fluid dynamics (very cool demo that's hard to capture in text)

#PyTorchCon

  • Copy link
  • Flag this comment
  • Block
phildini
@phildini@wandering.shop  ·  activity timestamp 2 months ago

Mathematical equations govern the world at all scales, and there's a need for physical understanding, which is outside the language-based domain of modern LLMs

#PyTorchCon

phildini
@phildini@wandering.shop replied  ·  activity timestamp 2 months ago

We need new neural operators that can handle physical, mathematical domains better.

Using lower-res translations and models considered harmful! Think: Hurricanes

#PyTorchCon TorchCon

  • Copy link
  • Flag this comment
  • Block
Log in

bonfire.cafe

A space for Bonfire maintainers and contributors to communicate

bonfire.cafe: About · Code of conduct · Privacy · Users · Instances
Bonfire social · 1.0.1-alpha.8 no JS en
Automatic federation enabled
  • Explore
  • About
  • Members
  • Code of Conduct
Home
Login