a LLM cookbook, for building your own from scratch, all the way from gathering data to training a model
-
Updated
Jun 25, 2024 - Jupyter Notebook
a LLM cookbook, for building your own from scratch, all the way from gathering data to training a model
This repository features a custom-built decoder-only language model (LLM) with a total of 37 million parameters 🔥. I train the model to be able to ask question from a given context
Experimental project for AI and NLP based on Transformer Architecture
Implementation of the GPT-3 paper: Language Models are Few-Shot Learners
Generate caption on images using CNN Encoder- LSTM Decoder structure
Transformers Intuition
从零实现基础的Transformer的Decoerder-Only模型,并进行模型升级,构建专属于自己的LLM模型
Generative AI fine-tune and inference for sequence classification tasks
An explainable and simplified version of OLMo model
a dna sequence generation/classification using transformers
This project aims to simplify texts from research papers using advanced natural language processing (NLP) techniques, making them more accessible to a broader audience
An LLM based tool for generation of cheese advirtisements
Code and dataset used to train dialect adapters for decoder models.
Custom decoder Transformer that treats a patient's medical journey like a story told through diagnosis codes instead of words.
Using LLMs in huggingface for sentiment analysis, translation, summarization and extractive question answering
Decoder model for language modelling
Coding A Decoder Only Transformer Like ChatGPT From Scratch
Intent Detection API using BERT and Flask
On the Design and Performance of Machine Learning Based Error Correcting Decoders
A mini version of GPT implemented on shakespear using BPE
Add a description, image, and links to the decoder-model topic page so that developers can more easily learn about it.
To associate your repository with the decoder-model topic, visit your repo's landing page and select "manage topics."