tom-caozh Goto Github PK
Name: Zhang Cao
Type: User
Bio: Interested in KV Stores, Cache, Disaggregated Memory(RDMA and CXL) and LLM.
Twitter: tomcaottt
Location: China
Name: Zhang Cao
Type: User
Bio: Interested in KV Stores, Cache, Disaggregated Memory(RDMA and CXL) and LLM.
Twitter: tomcaottt
Location: China
my solutions to CMU15-213 (updating)
my sections and homeworks of the standford class
to record my study in cs50_ai class
Curve is a high-performance, lightweight-operation, cloud-native open source distributed storage system. Curve can be applied to: 1) mainstream cloud-native infrastructure platforms OpenStack and Kubernetes; 2) high-performance storage for cloud-native databases; 3) cloud storage middleware using S3-compatible object storage as a data storage.
Contain some materials about CXL.
This is the implementation repository of our SOSP'23 paper: Ditto: An Elastic and Adaptive Memory-Disaggregated Caching System.
to stroe my config files
Running large language models on a single GPU for throughput-oriented scenarios.
Record my daily process when learning os-comp2022-winter
my solutions to some leetcode problems
To record some notes when I read the leveldb source code
Unify Efficient Fine-Tuning of 100+ LLMs
LLM inference in C/C++
Memkind is an easy-to-use, general-purpose allocator which helps to fully utilize various kinds of memory available in the system, including DRAM, NVDIMM, and HBM
to record my study of mit 6.824
OpenDAL: Access data freely, painlessly, and efficiently
Keep track of the papers I have read and to be read
High-speed Large Language Model Serving on PCs with Consumer-grade GPUs
To record some notes when I read the rocksdb source code
implement a rpc framework using golang, just for exercise
CLI tool for spawning and running containers according to the OCI specification
My rust study based on the cs110l course
NVIDIA® TensorRT™, an SDK for high-performance deep learning inference, includes a deep learning inference optimizer and runtime that delivers low latency and high throughput for inference applications.
TensorRT-LLM provides users with an easy-to-use Python API to define Large Language Models (LLMs) and build TensorRT engines that contain state-of-the-art optimizations to perform inference efficiently on NVIDIA GPUs. TensorRT-LLM also contains components to create Python and C++ runtimes that execute those TensorRT engines.
A declarative, efficient, and flexible JavaScript library for building user interfaces.
🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
An Open Source Machine Learning Framework for Everyone
The Web framework for perfectionists with deadlines.
A PHP framework for web artisans
Bring data to life with SVG, Canvas and HTML. 📊📈🎉
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
Some thing interesting about web. New door for the world.
A server is a program made to process requests and deliver data to clients.
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
Some thing interesting about visualization, use data art
Some thing interesting about game, make everyone happy.
We are working to build community through open source technology. NB: members must have two-factor auth.
Open source projects and samples from Microsoft.
Google ❤️ Open Source for everyone.
Alibaba Open Source for everyone
Data-Driven Documents codes.
China tencent open source team.