kimhyeonwoo's repositories
brave-tech-interview
🙋 핵심을 질문하다. 그리고 용감하게 대답하다. 국내 IT기업부터 실리콘밸리까지 "현직자가 해설해주는 기술면접"
IEEE-CIS-Fraud-Detection
캐글 메달 따기(가제)의 IEEE-CIS Fraud Detection 코드가 담긴 레포지토리입니다.
arithmetic
Code to reproduce "Transformers Can Do Arithmetic with the Right Embeddings", McLeish et al (2024)
DeepSeek-Coder
DeepSeek Coder: Let the Code Write Itself
DeepSpeed
DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.
DeepSpeedExamples
Example models using DeepSpeed
EHRSQL
EHRSQL: A Practical Text-to-SQL Benchmark for Electronic Health Records
gsm8k-ScRel
Codes and Data for Scaling Relationship on Learning Mathematical Reasoning with Large Language Models
huggingface-tutorial
트랜스포머를 활용한 자연어처리를 공부한 레포지토리입니다.
llm-meetup
LINER LLM Meetup archive
Megatron-LLaMA
Best practice for training LLaMA models in Megatron-LM
R-Tuning
Source code for the NAACL 2024 paper entitled "R-Tuning: Instructing Large Language Models to Say 'I Don't Know'"
repsys
Framework for developing and analyzing recommendation systems.
summarize-from-feedback
Code for "Learning to summarize from human feedback"
TensorRT-LLM
TensorRT-LLM provides users with an easy-to-use Python API to define Large Language Models (LLMs) and build TensorRT engines that contain state-of-the-art optimizations to perform inference efficiently on NVIDIA GPUs. TensorRT-LLM also contains components to create Python and C++ runtimes that execute those TensorRT engines.
ToRA
ToRA is a series of Tool-integrated Reasoning LLM Agents designed to solve challenging mathematical reasoning problems by interacting with tools.
Transformers-Tutorials
This repository contains demos I made with the Transformers library by HuggingFace.
TRON
⚡️ Implementation of TRON: Transformer Recommender using Optimized Negative-sampling, accepted at ACM RecSys 2023.
yacup_recsys_2022
1st place solution