blog_posts: 0f62568f1a
This data as json
id | createdDate | title | link | postExcerpt | featuredImageUrl | hash | contributors | modifiedDate | displayDate |
---|---|---|---|---|---|---|---|---|---|
blog-posts#34-72594 | 2024-03-18 21:25:12 | Optimize price-performance of LLM inference on NVIDIA GPUs using the Amazon SageMaker integration with NVIDIA NIM Microservices | https://aws.amazon.com/blogs/machine-learning/optimize-price-performance-of-llm-inference-on-nvidia-gpus-using-the-amazon-sagemaker-integration-with-nvidia-nim-microservices/ | NVIDIA NIM microservices now integrate with Amazon SageMaker, allowing you to deploy industry-leading large language models (LLMs) and optimize model performance and cost. You can deploy state-of-the-art LLMs in minutes instead of days using technologies such as NVIDIA TensorRT, NVIDIA TensorRT-LLM, and NVIDIA Triton Inference Server on NVIDIA accelerated instances hosted by SageMaker. NIM, part [...] | https://d2908q01vomqb2.cloudfront.net/f1f836cb4ea6efb2a0b1b99f41ad8b103eff4b59/2024/03/18/featured-images-ML-16498-300x169.jpg | 0f62568f1a | James Park, Eliuth Triana, Jiahong Liu, Kshitiz Gupta, Harish Tummalacherla, Nikhil Kulkarni, Qing Lan, Saurabh Trikande | 2024-03-18 21:25:12 | 18 Mar 2024 |
Links from other tables
- 5 rows from blog_post_hash in blog_post_tags