- 👋 Hi, I’m @hyuenmin-choi master student in KAIST(Korea Advanced Institution of Science and Technology)
- 👀 I’m interested in computer architecture especially in system for AI!
- 🌱 I’m currently researching Large-Scale LLM inference serving system optimization techniques spanning from hardware accelerators to software design.
- 📫 Contact me -> wisestone1213@gmail.com or hmchoi@casys.kaist.ac.kr
Popular repositories Loading
-
Orca_reproduce.code
Orca_reproduce.code PublicMimicing ORCA[OSDI'22] serving system using triton base implemetation
-
DRAMSim2
DRAMSim2 PublicForked from umd-memsys/DRAMSim2
DRAMSim2: A cycle accurate DRAM simulator
C++ 1
-
Orca-triton-server
Orca-triton-server PublicForked from triton-inference-server/server
The Triton Inference Server provides an optimized cloud and edge inferencing solution.
Python 1
-
ORCA-triton-core
ORCA-triton-core PublicForked from triton-inference-server/core
The core library and APIs implementing the Triton Inference Server.
C++ 1
-
splitwise-sim
splitwise-sim PublicForked from mutinifni/splitwise-sim
LLM serving cluster simulator
Jupyter Notebook 1
-
If the problem persists, check the GitHub status page or contact support.
