Mr-Nineteen

Mr-Nineteen

Geek Repo

Location:Shanghai

Github PK Tool:Github PK Tool

Mr-Nineteen's repositories

RecSysPapers

推荐/广告/搜索领域工业界经典以及最前沿论文集合。A collection of industry classics and cutting-edge papers in the field of recommendation/advertising/search.

Language:PythonLicense:BSD-2-ClauseStargazers:1Issues:0Issues:0
Language:Jupyter NotebookStargazers:0Issues:0Issues:0

baichuan-Dynamic-NTK-ALiBi

百川Dynamic NTK-ALiBi的代码实现:无需微调即可推理更长文本

Language:PythonStargazers:0Issues:0Issues:0

ChatGLM2-6B

ChatGLM2-6B: An Open Bilingual Chat LLM | 开源双语对话语言模型

Language:PythonLicense:NOASSERTIONStargazers:0Issues:0Issues:0

Chinese-LlaMA2

Repo for adapting Meta LlaMA2 in Chinese! META最新发布的LlaMA2的汉化版! (完全开源可商用)

Language:PythonStargazers:0Issues:0Issues:0

CUDA-Programming

Sample codes for my CUDA programming book

Language:CudaLicense:GPL-3.0Stargazers:0Issues:0Issues:0

cuda-samples

Samples for CUDA Developers which demonstrates features in CUDA Toolkit

Language:CLicense:NOASSERTIONStargazers:0Issues:0Issues:0

CUDALibrarySamples

CUDA Library Samples

Language:CudaLicense:NOASSERTIONStargazers:0Issues:0Issues:0

diffusers

🤗 Diffusers: State-of-the-art diffusion models for image and audio generation in PyTorch

Language:PythonLicense:Apache-2.0Stargazers:0Issues:0Issues:0

flash-attention

Fast and memory-efficient exact attention

Language:PythonLicense:BSD-3-ClauseStargazers:0Issues:0Issues:0

generative-recommenders

Repository hosting code used to reproduce results in "Actions Speak Louder than Words: Trillion-Parameter Sequential Transducers for Generative Recommendations" (https://arxiv.org/abs/2402.17152).

Language:PythonLicense:Apache-2.0Stargazers:0Issues:0Issues:0

grok-1

Grok open release

Language:PythonLicense:Apache-2.0Stargazers:0Issues:0Issues:0

HierarchicalKV

HierarchicalKV is a part of NVIDIA Merlin and provides hierarchical key-value storage to meet RecSys requirements. The key capability of Merlin-KV is to store key-value feature-embeddings on high-bandwidth memory (HBM) of GPUs and in host memory. It also can be used as a generic key-value storage.

Language:CudaLicense:Apache-2.0Stargazers:0Issues:0Issues:0

how-to-optim-algorithm-in-cuda

how to optimize some algorithm in cuda.

Language:CudaStargazers:0Issues:0Issues:0

iTransformer

Official implementation for "iTransformer: Inverted Transformers Are Effective for Time Series Forecasting" (ICLR 2024 Spotlight), https://openreview.net/forum?id=JePfAI8fah

Language:PythonLicense:MITStargazers:0Issues:0Issues:0

llama

Inference code for LLaMA models

Language:PythonLicense:NOASSERTIONStargazers:0Issues:0Issues:0

LLaMA-Efficient-Tuning

Fine-tuning LLaMA with PEFT (PT+SFT+RLHF with QLoRA)

Language:PythonLicense:Apache-2.0Stargazers:0Issues:0Issues:0

llama-recipes

Examples and recipes for Llama 2 model

Language:PythonLicense:NOASSERTIONStargazers:0Issues:0Issues:0

llm-action

本项目旨在分享大模型相关技术原理以及实战经验。

Language:PythonLicense:Apache-2.0Stargazers:0Issues:0Issues:0

nvcomp

Repository for nvCOMP docs and examples. nvCOMP is a library for fast lossless compression/decompression on the GPU that can be downloaded from https://developer.nvidia.com/nvcomp.

Language:C++License:NOASSERTIONStargazers:0Issues:0Issues:0

onnx

Open standard for machine learning interoperability

Language:PythonLicense:Apache-2.0Stargazers:0Issues:0Issues:0

onnx-simplifier

Simplify your onnx model

Language:C++License:Apache-2.0Stargazers:0Issues:0Issues:0

onnxruntime-inference-examples

Examples for using ONNX Runtime for machine learning inferencing.

Language:PythonLicense:MITStargazers:0Issues:0Issues:0

tensorrtx

Implementation of popular deep learning networks with TensorRT network definition API

Language:C++License:MITStargazers:0Issues:0Issues:0

Time-Series-Library

A Library for Advanced Deep Time Series Models.

Language:ShellLicense:MITStargazers:0Issues:0Issues:0
Language:PythonStargazers:0Issues:1Issues:0

transformers

🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.

Language:PythonLicense:Apache-2.0Stargazers:0Issues:0Issues:0

transformers-stream-generator

This is a text generation method which returns a generator, streaming out each token in real-time during inference, based on Huggingface/Transformers.

Language:PythonLicense:MITStargazers:0Issues:0Issues:0

Transformers-Tutorials

This repository contains demos I made with the Transformers library by HuggingFace.

Language:Jupyter NotebookLicense:MITStargazers:0Issues:0Issues:0

vllm

A high-throughput and memory-efficient inference and serving engine for LLMs

Language:PythonLicense:Apache-2.0Stargazers:0Issues:0Issues:0