Hello, I'm Teknium1 π
I'm a Python Programmer, AI Enthusiast, and a Co-founder of NousResearch.
My work primarily involves AI and Data Engineering, contributing primarily by releasing open source Large Language Model (LLMs) and datasets.
I've contributed significantly to the development of several opensource LLMs under Nous Research's huggingface organization.
Here are a couple of them:
- Nous-Hermes-2-Mixtral-8x7B - Nous' latest most powerful LLM yet.
- Nous-Hermes-2-Yi-34B - Hermes 2 on Yi-34B.
- Nous-Hermes-Llama2-13b - A Hermes model built on llama 1 and llama 2.
- GPT4-x-Vicuna-13b - A Vicuna model built on GPT-4.
On my personal huggingface, Teknium, I have released several models, including my work on Replit-3b Model & OpenHermes:
- DataForge - Economics - A dataset built by my data synthesis pipeline (not public), DataForge
- OpenHermes 2.5 Mistral 7B - Most powerful Open Hermes, with much improved coding skills than OpenHermes 2
- OpenHermes 2 Mistral 7B - Version 2 of the Open Hermes series.
- OpenHermes 13B - An Open Sourced version of Nous-Hermes!
- OpenHermes Dataset - The publicly available version of Hermes' dataset.
- Replit-Instruct 3B - This model doubled the code performance of the LLM.
I've been part of several intriguing projects on GitHub. Here are a few of them:
- LLM-Benchmark-Logs - A repository full of benchmarks I've done on various LLMs, originally inside of Nous' discord but it became too disorganized, so now lives on Github.
- LLM-Logbook - A temporary project that became too expensive to do, collection of responses for 100 random crowdsourced prompts to various LLMs.
- GPTeacher - A collection of modular datasets generated by GPT-4, for training LLMs.
- RawTransform - A repository of prompts and Python scripts for intelligent transformation of raw text into diverse formats.
- stanford_alpaca-replit - Modified Stanford-Alpaca Trainer for Training Replit's Code Model.
- alpaca-roleplay-discordbot - An LLM discord bot that roleplays!
- alpaca-discord - A Simple Discord Bot for the Alpaca LLM.
Have worked on researching, planning ablations, and cleaning/filtering the dataset for:
- StableBeluga/Free Willy 2 - Orca replication on 70b Llama-2
- StableBeluga/Free-Willy-1 - Orca replication on 65b Llama-1
Both are 10% Orca replications trained on Llama-1 and Llama-2 70B. Also working on domain expert knowledge and task distillation.
Working with the Open Orca team on data cleaning, networking, ablations, and more:
- Open Orca HuggingFace Repo - An Open Sourced Orca paper replication
- Twitter: https://twitter.com/Teknium1
- Discord: Teknium