📚 Auto-publish: Add/update 3 blog posts
All checks were successful
Hugo Publish CI / build-and-deploy (push) Successful in 11s

Generated on: Sat Oct  4 17:44:47 UTC 2025
Source: md-personal repository
This commit is contained in:
Automated Publisher
2025-10-04 17:44:47 +00:00
parent 85e0d053b7
commit 2f73eaed9a
3 changed files with 2 additions and 2 deletions

View File

@@ -55,8 +55,7 @@ To understand where performance hits its ceiling, I applied roofline analysis—
The roofline model works by comparing a workload's operational intensity (how many calculations you do per byte of data moved) against the device's balance point. If your operational intensity is too low, you're bottlenecked by memory bandwidth—and as we'll see, that's exactly what happens with LLM inference.
![S3 File](/images/benchmarking-llms-on-jetson-orin-nano/388f43c3f800483aae5ea487e8f45922.png)
![S3 File](/images/benchmarking-llms-on-jetson-orin-nano/16d64bdc9cf14b05b7c40c4718b8091b.png)
## The Results: Speed and Efficiency