InAWS in Plain EnglishbyRam VegirajuDeploying Transformers ONNX Models on Amazon SageMakerAchieve High Scale Performance Utilizing Triton Inference Server With SageMaker Real-Time InferenceMar 13Mar 13
InTowards Data SciencebyRam VegirajuAn Introduction To Fine-Tuning Pre-Trained Transformers ModelsSimplified utilizing the HuggingFace trainer objectFeb 17Feb 17
InTowards Data SciencebyRam VegirajuDeploying Large Language Models With HuggingFace TGIAnother way to efficiently host and scale your LLMs with Amazon SageMakerJul 14, 20231Jul 14, 20231
InTowards Data SciencebyRam VegirajuDeploying LLMs On Amazon SageMaker With DJL ServingDeploy BART on Amazon SageMaker Real-Time InferenceJun 7, 2023Jun 7, 2023
InTowards Data SciencebyRam VegirajuHuggingFace Inference EndpointsRapid production-grade deployment of Transformers modelsDec 23, 20221Dec 23, 20221
InTowards Data SciencebyRam VegirajuDeploy HuggingFace Models on Amazon SageMaker under 10 MinutesUse the HuggingFace Hub to Deploy NLP Models at ScaleOct 6, 20211Oct 6, 20211
InTowards Data SciencebyRam VegirajuHost ML Apps with HuggingFace SpacesExploring the New HuggingFace ReleaseNov 7, 2021Nov 7, 2021