ASWATH RAM's repositories
ec2instances.info
Amazon EC2 instance comparison site
amazon-bedrock-workshop
This is a workshop designed for Amazon Bedrock a foundational model service.
American-Sign-Lnaguage-Detection
American Sign language detector using mediapipe
Profile
Profile Page
CC1.2-Project-Group5
Public repo of our HTML/CSS project for Group 5
Corrective-RAG
Implementation of Corrective RAG using LangChain and LangGraph.
CRAG
Corrective Retrieval Augmented Generation
deep-learning-containers
AWS Deep Learning Containers (DLCs) are a set of Docker images for training and serving models in TensorFlow, TensorFlow 2, PyTorch, and MXNet.
DeepSpeed
DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.
djl
An Engine-Agnostic Deep Learning Framework in Java
EC2-timeline
A historical overview of AWS EC2 instance releases -
ficto
Ficto is a Python package that allows you to effortlessly generate realistic dummy data in CSV or JSON format.
foundation-model-benchmarking-tool
Foundation model benchmarking tool. Run any model on Amazon SageMaker and benchmark for performance across instance type and serving stack options.
llama-recipes
Scripts for fine-tuning Llama2 with composable FSDP & PEFT methods to cover single/multi-node GPUs. Supports default & custom datasets for applications such as summarization & question answering. Supporting a number of candid inference solutions such as HF TGI, VLLM for local or cloud deployment.Demo apps to showcase Llama2 for WhatsApp & Messenger
Sagemaker-BYOC
Sagemaker BYOC with neuron Compiler
sagemaker-debugger
Amazon SageMaker Debugger provides functionality to save tensors during training of machine learning jobs and analyze those tensors
sagemaker-explaining-credit-decisions
Amazon SageMaker Solution for explaining credit decisions.
SageMaker-JupyterLab-Images
Images that work with sageMaker JupyterLab
sagemaker-python-sdk
A library for training and deploying machine learning models on Amazon SageMaker
TensorRT-LLM
TensorRT-LLM provides users with an easy-to-use Python API to define Large Language Models (LLMs) and build TensorRT engines that contain state-of-the-art optimizations to perform inference efficiently on NVIDIA GPUs. TensorRT-LLM also contains components to create Python and C++ runtimes that execute those TensorRT engines.
text-generation-inference
Large Language Model Text Generation Inference
visrec-djl
Implementation of JSR381's API using DJL.