Júlio César's Projects
Como criar uma linguagem de programação simples do zero, usando somente o ANTLR4, o Windows e o ChatGPT3. Tutorial completo.
33B Chinese LLM, DPO QLORA, 100K context, AirLLM 70B inference with single 4GB GPU
ANTLR (ANother Tool for Language Recognition) is a powerful parser generator for reading, processing, executing, or translating structured text or binary files.
An experimental open-source attempt to make GPT-4 fully autonomous.
Go ahead and axolotl questions
A bagel, with everything.
CodeBERT
CodeQwen1.5 is the code version of Qwen, the large language model series developed by Qwen team, Alibaba Cloud.
Pretrained Language Models for Source code
Framework for orchestrating role-playing, autonomous AI agents. By fostering collaborative intelligence, CrewAI empowers agents to work together seamlessly, tackling complex tasks.
Suite of tools for deploying and training deep learning models using the JVM. Highlights include model import for keras, tensorflow, and onnx/pytorch, a modular and tiny c++ library for running math code and a java based math library on top of the core c++ library. Also includes samediff: a pytorch/tensorflow like library for running deep learning
DeepSeek Coder: Let the Code Write Itself
DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.
Devika is an Agentic AI Software Engineer that can understand high-level human instructions, break them down into steps, research relevant information, and write code to achieve the given objective. Devika aims to be a competitive open-source alternative to Devin by Cognition AI.
Evals is a framework for evaluating LLMs and LLM systems, and an open-source registry of benchmarks.
🤗 Evaluate: A library for easily evaluating machine learning models and datasets.
Fine-tuning Large Language Models (LLMs) is a crucial step in adapting these powerful models to specific tasks or domains. In this seminar code tutorial, we will explore how to perform fine-tuning using QLoRA (Quantized LoRA), a memory-efficient iteration of LoRA (Low-Rank Adaptation), for parameter-efficient fine-tuning.
Tensor library for machine learning
An implementation of model parallel GPT-2 and GPT-3-style models using the mesh-tensorflow library.
An implementation of model parallel autoregressive transformers on GPUs, based on the DeepSpeed library.
Code for the paper "Evaluating Large Language Models Trained on Code"
ANTLR as a libray for JVM based languages
Unify Efficient Fine-Tuning of 100+ LLMs
Preset JSON files to be used with LM studio
Lua for Windows is a 'batteries included environment' for the Lua scripting language on Windows. NOTICE: Looking for maintainer.
Magicoder: Source Code Is All You Need
MobiLlama : Small Language Model tailored for edge devices
OpenCodeInterpreter is a suite of open-source code generation systems aimed at bridging the gap between large language models and sophisticated proprietary systems like the GPT-4 Code Interpreter. It significantly enhances code generation capabilities by integrating execution and iterative refinement functionalities.