Pinned Repositories
alpaca-lora
Instruct-tune LLaMA on consumer hardware
attention_sinks
Extend existing LLMs way beyond the original training length with constant memory usage, and without retraining
bert4keras
keras implement of transformers for humans
ChatGLM-Efficient-Tuning
Fine-tuning ChatGLM-6B with PEFT | 基于 PEFT 的高效 ChatGLM 微调
ChatGLM2-6B
ChatGLM2-6B: An Open Bilingual Chat LLM | 开源双语对话语言模型
ColossalAI
Making big AI models cheaper, easier, and more scalable
contrastive-htc
This repository implements a contrastive learning model for hierarchical text classification. This work has been accepted as the long paper "Incorporating Hierarchy into Text Encoder: a Contrastive Learning Approach for Hierarchical Text Classification" in ACL 2022.
customer_need_explore
DeepInterestNetwork
DIN
examples
All the code examples I use in my videos
felixstander's Repositories
felixstander/alpaca-lora
Instruct-tune LLaMA on consumer hardware
felixstander/attention_sinks
Extend existing LLMs way beyond the original training length with constant memory usage, and without retraining
felixstander/bert4keras
keras implement of transformers for humans
felixstander/ChatGLM-Efficient-Tuning
Fine-tuning ChatGLM-6B with PEFT | 基于 PEFT 的高效 ChatGLM 微调
felixstander/ChatGLM2-6B
ChatGLM2-6B: An Open Bilingual Chat LLM | 开源双语对话语言模型
felixstander/ColossalAI
Making big AI models cheaper, easier, and more scalable
felixstander/contrastive-htc
This repository implements a contrastive learning model for hierarchical text classification. This work has been accepted as the long paper "Incorporating Hierarchy into Text Encoder: a Contrastive Learning Approach for Hierarchical Text Classification" in ACL 2022.
felixstander/customer_need_explore
felixstander/examples
All the code examples I use in my videos
felixstander/Finetune-ChatGLM2-6B
ChatGLM2-6B 全参数微调,支持多轮对话的高效微调。
felixstander/FinGLM
felixstander/Firefly
Firefly(流萤): 中文对话式大语言模型(全量微调+QLoRA),支持微调Baichuan2、CodeLlama、Llma2、Llama、Qwen、Baichuan、ChatGLM2、InternLM、Ziya、Bloom等大模型
felixstander/GPT2-chitchat
GPT2 for Chinese chitchat/用于中文闲聊的GPT2模型(实现了DialoGPT的MMI**)
felixstander/GrailQA
felixstander/guidance
A guidance language for controlling large language models.
felixstander/KnowPrompt
Code and datasets for the WWW2022 paper "KnowPrompt: Knowledge-aware Prompt-tuning with Synergistic Optimization for Relation Extraction."
felixstander/lit
The Language Interpretability Tool: Interactively analyze NLP models for model understanding in an extensible and framework agnostic interface.
felixstander/LLM_pratice
关于 LLM知识点
felixstander/lmdeploy
LMDeploy is a toolkit for compressing, deploying, and serving LLMs.
felixstander/NBA_chaser
跟踪每日 nba 热点,搭建 nba 问答机器人
felixstander/PaddleNLP
Easy-to-use and powerful NLP library with Awesome model zoo, supporting wide-range of NLP tasks from research to industrial applications, including Neural Search, Question Answering, Information Extraction and Sentiment Analysis end-to-end system.
felixstander/Qwen-7B
The official repo of Qwen-7B (通义千问-7B) chat & pretrained large language model proposed by Alibaba Cloud.
felixstander/rasat
The official implementation of the paper "RASAT: Integrating Relational Structures into Pretrained Seq2Seq Model for Text-to-SQL"(EMNLP 2022)
felixstander/self-chat
felixstander/self-rag
This includes the original implementation of SELF-RAG: Learning to Retrieve, Generate and Critique through self-reflection by Akari Asai, Zeqiu Wu, Yizhong Wang, Avirup Sil, and Hannaneh Hajishirzi.
felixstander/streaming-llm
Efficient Streaming Language Models with Attention Sinks
felixstander/street-fighter-bot
QA bot for answering street-fighter game questions
felixstander/text-generation-webui
A gradio web UI for running Large Language Models like GPT-J 6B, OPT, GALACTICA, LLaMA, and Pygmalion.
felixstander/transformers-bloom-inference
Fast Inference Solutions for BLOOM
felixstander/xtuner
XTuner is a toolkit for efficiently fine-tuning LLM