Difference between revisions of "Free Software Directory:Artificial Intelligence Team"
GrahamxReed (talk | contribs) m (added neat looking agnai project (multiple character chatbot)) |
GrahamxReed (talk | contribs) (Image generation descriptions; removed StabilityAI listings: generally not a fan of the multiple updatings to Open RAIL model licenses when it is unclear how free-as-in-freedom they are (better safe than sorry).) |
||
Line 82: | Line 82: | ||
| agnaistic | | agnaistic | ||
| [https://github.com/agnaistic/agnai#AGPL-3.0-1-ov-file AGPLv3] | | [https://github.com/agnaistic/agnai#AGPL-3.0-1-ov-file AGPLv3] | ||
− | | AI | + | | AI agnostic (multi-user and multi-bot) chat with fictional characters |
|- | |- | ||
! [https://github.com/turboderp/exllamav2 ExLlamaV2] | ! [https://github.com/turboderp/exllamav2 ExLlamaV2] | ||
Line 117: | Line 117: | ||
| SillyTavern | | SillyTavern | ||
| [https://github.com/SillyTavern/SillyTavern/blob/release/LICENSE AGPLv3] | | [https://github.com/SillyTavern/SillyTavern/blob/release/LICENSE AGPLv3] | ||
− | | LLM | + | | LLM frontend for power users |
|- | |- | ||
! [https://github.com/oobabooga/text-generation-webui Text Generation Web UI] | ! [https://github.com/oobabooga/text-generation-webui Text Generation Web UI] | ||
| oobabooga | | oobabooga | ||
| [https://github.com/oobabooga/text-generation-webui/blob/main/LICENSE AGPLv3] | | [https://github.com/oobabooga/text-generation-webui/blob/main/LICENSE AGPLv3] | ||
− | | A Gradio web UI for Large Language Models. Supports transformers, GPTQ, AWQ, EXL2, llama.cpp (GGUF), | + | | A Gradio web UI for Large Language Models. Supports transformers, GPTQ, AWQ, EXL2, llama.cpp (GGUF), LLaMA models |
|- | |- | ||
|} | |} | ||
Line 166: | Line 166: | ||
| lm-sys | | lm-sys | ||
| [https://github.com/lm-sys/FastChat/blob/main/LICENSE Apache-2.0] | | [https://github.com/lm-sys/FastChat/blob/main/LICENSE Apache-2.0] | ||
− | | An open platform for training, serving, and evaluating large language models | + | | An open platform for training, serving, and evaluating large language models |
|- | |- | ||
! [https://github.com/LAION-AI/Open-Assistant Open Assistant] | ! [https://github.com/LAION-AI/Open-Assistant Open Assistant] | ||
Line 176: | Line 176: | ||
| KillianLucas | | KillianLucas | ||
| [https://github.com/KillianLucas/open-interpreter/blob/main/LICENSE AGPLv3] | | [https://github.com/KillianLucas/open-interpreter/blob/main/LICENSE AGPLv3] | ||
− | | Lets LLMs run code (Python, Javascript, Shell, and more) locally | + | | Lets LLMs run code (Python, Javascript, Shell, and more) locally |
|- | |- | ||
|} | |} | ||
Line 191: | Line 191: | ||
== Images == | == Images == | ||
=== Image generation === | === Image generation === | ||
− | ==== Text to image ==== | + | ==== Text to image GUI ==== |
{| class="wikitable sortable" | {| class="wikitable sortable" | ||
|- | |- | ||
Line 197: | Line 197: | ||
! Credit | ! Credit | ||
! License | ! License | ||
+ | ! Description | ||
|- | |- | ||
! [https://github.com/comfyanonymous/ComfyUI ComfyUI] | ! [https://github.com/comfyanonymous/ComfyUI ComfyUI] | ||
| comfyanonymous | | comfyanonymous | ||
| [https://github.com/comfyanonymous/ComfyUI/blob/master/LICENSE GPLv3] | | [https://github.com/comfyanonymous/ComfyUI/blob/master/LICENSE GPLv3] | ||
+ | | Modular Stable Diffusion GUI, API, and backend with a graph/nodes interface | ||
+ | |- | ||
+ | ! [https://github.com/AUTOMATIC1111/stable-diffusion-webui Stable Diffusion WebUI] | ||
+ | | AUTOMATIC1111 | ||
+ | | [https://github.com/AUTOMATIC1111/stable-diffusion-webui/blob/master/LICENSE.txt AGPLv3] | ||
+ | | Stable Diffusion web UI | ||
+ | |- | ||
+ | |} | ||
+ | ==== Additional control libraries ==== | ||
+ | {| class="wikitable sortable" | ||
+ | |- | ||
+ | ! Project | ||
+ | ! Credit | ||
+ | ! License | ||
+ | ! Description | ||
|- | |- | ||
! [https://github.com/lllyasviel/ControlNet ControlNet] | ! [https://github.com/lllyasviel/ControlNet ControlNet] | ||
| lllyasviel | | lllyasviel | ||
| [https://github.com/lllyasviel/ControlNet/blob/main/LICENSE Apache-2.0] | | [https://github.com/lllyasviel/ControlNet/blob/main/LICENSE Apache-2.0] | ||
+ | | Adding conditional control to text-to-image diffusion models | ||
+ | |- | ||
+ | ! [https://github.com/huggingface/diffusers Diffusers] | ||
+ | | huggingface | ||
+ | | [https://github.com/huggingface/diffusers/blob/main/LICENSE Apache-2.0] | ||
+ | | State-of-the-art diffusion models for image and audio generation in PyTorch | ||
+ | |- | ||
+ | |} | ||
+ | ==== Legacy ==== | ||
+ | {| class="wikitable sortable" | ||
+ | |- | ||
+ | ! Project | ||
+ | ! Credit | ||
+ | ! License | ||
+ | ! Description | ||
|- | |- | ||
! [https://github.com/borisdayma/dalle-mini/ DALL-E Mini] | ! [https://github.com/borisdayma/dalle-mini/ DALL-E Mini] | ||
| borisdayma (Boris Dayma) | | borisdayma (Boris Dayma) | ||
| [https://github.com/borisdayma/dalle-mini/blob/main/LICENSE Apache 2.0] | | [https://github.com/borisdayma/dalle-mini/blob/main/LICENSE Apache 2.0] | ||
− | | | + | | Generate images from a text prompt |
− | |||
− | |||
− | |||
|- | |- | ||
! [https://github.com/anishathalye/neural-style neural-style] | ! [https://github.com/anishathalye/neural-style neural-style] | ||
| anishathalye | | anishathalye | ||
| [https://github.com/anishathalye/neural-style/blob/master/LICENSE.txt GPLv3] | | [https://github.com/anishathalye/neural-style/blob/master/LICENSE.txt GPLv3] | ||
− | | | + | | An implementation of neural style in TensorFlow |
− | |||
− | |||
− | |||
− | |||
− | |||
− | |||
− | |||
|- | |- | ||
|} | |} | ||
− | + | ===Videos=== | |
Image generation techniques create pictures from noise estimations. This noise shows up as artifacts and hampers temporal stability for objects. These projects tackle that issue. | Image generation techniques create pictures from noise estimations. This noise shows up as artifacts and hampers temporal stability for objects. These projects tackle that issue. | ||
{| class="wikitable sortable" | {| class="wikitable sortable" | ||
Line 235: | Line 256: | ||
! Credit | ! Credit | ||
! License | ! License | ||
+ | ! Description | ||
|- | |- | ||
! [https://github.com/google-research/frame-interpolation FILM: Frame Interpolation for Large Motion] | ! [https://github.com/google-research/frame-interpolation FILM: Frame Interpolation for Large Motion] | ||
| Google Research | | Google Research | ||
| [https://github.com/google-research/frame-interpolation#Apache-2.0-1-ov-file Apache 2.0] | | [https://github.com/google-research/frame-interpolation#Apache-2.0-1-ov-file Apache 2.0] | ||
+ | | A unified single-network approach to frame interpolation that doesn't use additional pre-trained networks | ||
|- | |- | ||
− | ! [https://github.com/magic-research/magic-animate MagicAnimate | + | ! [https://github.com/magic-research/magic-animate MagicAnimate] |
| MagIC Research | | MagIC Research | ||
| [https://github.com/magic-research/magic-animate#BSD-3-Clause-1-ov-file BSD-3-Clause license] | | [https://github.com/magic-research/magic-animate#BSD-3-Clause-1-ov-file BSD-3-Clause license] | ||
+ | | Temporally consistent human image animation using a diffusion model | ||
|- | |- | ||
− | ! [https://github.com/TencentARC/MotionCtrl MotionCtrl | + | ! [https://github.com/TencentARC/MotionCtrl MotionCtrl] |
| ARC Lab, Tencent PCG | | ARC Lab, Tencent PCG | ||
| [https://github.com/TencentARC/MotionCtrl#Apache-2.0-1-ov-file Apache 2.0] | | [https://github.com/TencentARC/MotionCtrl#Apache-2.0-1-ov-file Apache 2.0] | ||
− | | | + | | A unified and flexible motion controller for video generation |
− | |||
− | |||
− | |||
|- | |- | ||
! [https://github.com/CiaraStrawberry/TemporalKit TemporalKit] | ! [https://github.com/CiaraStrawberry/TemporalKit TemporalKit] | ||
| CiaraStrawberry (Ciara Rowles) | | CiaraStrawberry (Ciara Rowles) | ||
| [https://github.com/CiaraStrawberry/TemporalKit?tab=GPL-3.0-1-ov-file#readme GPLv3] | | [https://github.com/CiaraStrawberry/TemporalKit?tab=GPL-3.0-1-ov-file#readme GPLv3] | ||
+ | | An all in one solution for adding temporal stability to a Stable Diffusion render via an automatic1111 extension | ||
|- | |- | ||
! [https://github.com/camenduru/text-to-video-synthesis-colab Text To Video Synthesis Colab] | ! [https://github.com/camenduru/text-to-video-synthesis-colab Text To Video Synthesis Colab] | ||
| camenduru | | camenduru | ||
| [https://github.com/camenduru/text-to-video-synthesis-colab/blob/main/LICENSE The Unlicense] | | [https://github.com/camenduru/text-to-video-synthesis-colab/blob/main/LICENSE The Unlicense] | ||
+ | | A text-to-video synthesis model that evolves from a text-to-image synthesis model | ||
|- | |- | ||
! [https://github.com/yoyo-nb/Thin-Plate-Spline-Motion-Model Thin-Plate Spline Motion Model for Image Animation] | ! [https://github.com/yoyo-nb/Thin-Plate-Spline-Motion-Model Thin-Plate Spline Motion Model for Image Animation] | ||
| yoyo-nb | | yoyo-nb | ||
| [https://github.com/yoyo-nb/Thin-Plate-Spline-Motion-Model/blob/main/LICENSE MIT] | | [https://github.com/yoyo-nb/Thin-Plate-Spline-Motion-Model/blob/main/LICENSE MIT] | ||
+ | | Animates a static object in a source image according to a driving video | ||
|- | |- | ||
|} | |} |
Revision as of 01:55, 2 March 2024
The Artificial Intelligence Project Team gathers free software resources regarding machine learning / artificial intelligence.
Group info | User info | |||||
---|---|---|---|---|---|---|
User | Role | Reference | Real name | libera.chat nick | Time zone | Title |
David_Hedlund | Coordinator | David Hedlund | David_Hedlund | Europe/Stockholm | ||
GrahamxReed | Collaborator | Graham Reed | Graham_Reed | America/New_York | ||
Mertgor | Observer | Mert Gör | hwpplayer1 | Europe/Istanbul | ||
Mmcmahon | Team captain | Michael McMahon | thomzane | America/New_York | FSF Systems Administrator |
Truthfulness
- US NIST AI Risk Management Framework Playbook
- A science paper benchmark: TruthfulQA: Measuring How Models Mimic Human Falsehoods
Transformers
- Package manager and player: Transformers - https://github.com/huggingface/transformers/
- Original science paper: Attention Is All You Need
Text
Grammar
Translation
Text generation
Front ends
Project | Credit | License | Description |
---|---|---|---|
Agnai | agnaistic | AGPLv3 | AI agnostic (multi-user and multi-bot) chat with fictional characters |
ExLlamaV2 | turboderp | MIT License | A fast inference library for running LLMs locally on modern consumer-class GPUs |
ExLlamaV2 WebUI | turboderp | MIT License | |
GPT4All | Nomic AI | MIT | Run open-source LLMs anywhere |
KoboldAI | KoboldAI | AGPLv3 | A browser-based front-end for AI-assisted writing with multiple local & remote AI models |
koboldcpp | LostRuins | AGPLv3 | A simple one-file way to run various GGML and GGUF models with KoboldAI's UI |
Serge | serge-chat | MIT | A web interface for chatting with Alpaca through llama.cpp. Fully dockerized, with an easy to use API. |
SillyTavern | SillyTavern | AGPLv3 | LLM frontend for power users |
Text Generation Web UI | oobabooga | AGPLv3 | A Gradio web UI for Large Language Models. Supports transformers, GPTQ, AWQ, EXL2, llama.cpp (GGUF), LLaMA models |
Models
Project | Credit | License | Description |
---|---|---|---|
llama.cpp | ggerganov (Georgi Gerganov) | MIT | Port of Facebook's LLaMA model in C/C++ |
Mistral | Mistral AI | Apache 2.0 | Mistral 7B significantly outperforms Llama 2 13B on all metrics, and is on par with Llama 34B. |
Mixtral | Mistral AI | Apache 2.0 | Mixtral outperforms Llama 2 70B on most benchmarks with 6x faster inference. |
Other
Project | Credit | License | Description |
---|---|---|---|
GPT-NeoX | EleutherAI | Apache 2.0 | EleutherAI's library for training large-scale language models on GPUs |
FastChat | lm-sys | Apache-2.0 | An open platform for training, serving, and evaluating large language models |
Open Assistant | LAION-AI | Apache 2.0 | A chat-based assistant that understands tasks, can interact with third-party systems, and retrieve information dynamically to do so |
Open Interpreter | KillianLucas | AGPLv3 | Lets LLMs run code (Python, Javascript, Shell, and more) locally |
Honorary mention: LLaMA 1 had an AGPLv3 license.
Code generation
This concept is controversial. See the FSF's other writing on this topic.
- CodeGen by Salesforce FSD | BSD 3-Clause "New" or "Revised" License
- TurboPilot by ravenscroftj (James Ravenscroft) FSD | BSD 3-Clause "New" or "Revised" License
- CodeGen2 by Salesforce FSD | Apache-2.0
Images
Image generation
Text to image GUI
Project | Credit | License | Description |
---|---|---|---|
ComfyUI | comfyanonymous | GPLv3 | Modular Stable Diffusion GUI, API, and backend with a graph/nodes interface |
Stable Diffusion WebUI | AUTOMATIC1111 | AGPLv3 | Stable Diffusion web UI |
Additional control libraries
Project | Credit | License | Description |
---|---|---|---|
ControlNet | lllyasviel | Apache-2.0 | Adding conditional control to text-to-image diffusion models |
Diffusers | huggingface | Apache-2.0 | State-of-the-art diffusion models for image and audio generation in PyTorch |
Legacy
Project | Credit | License | Description |
---|---|---|---|
DALL-E Mini | borisdayma (Boris Dayma) | Apache 2.0 | Generate images from a text prompt |
neural-style | anishathalye | GPLv3 | An implementation of neural style in TensorFlow |
Videos
Image generation techniques create pictures from noise estimations. This noise shows up as artifacts and hampers temporal stability for objects. These projects tackle that issue.
Project | Credit | License | Description |
---|---|---|---|
FILM: Frame Interpolation for Large Motion | Google Research | Apache 2.0 | A unified single-network approach to frame interpolation that doesn't use additional pre-trained networks |
MagicAnimate | MagIC Research | BSD-3-Clause license | Temporally consistent human image animation using a diffusion model |
MotionCtrl | ARC Lab, Tencent PCG | Apache 2.0 | A unified and flexible motion controller for video generation |
TemporalKit | CiaraStrawberry (Ciara Rowles) | GPLv3 | An all in one solution for adding temporal stability to a Stable Diffusion render via an automatic1111 extension |
Text To Video Synthesis Colab | camenduru | The Unlicense | A text-to-video synthesis model that evolves from a text-to-image synthesis model |
Thin-Plate Spline Motion Model for Image Animation | yoyo-nb | MIT | Animates a static object in a source image according to a driving video |
- MiniGPT-4 by Vision-CAIR | BSD 3-Clause "New" or "Revised" License & BSD 3-Clause License
- LLaVA by Haotian Liu | Apache License 2.0
3D modeling
Audio
Natural language processing (NLP)
Transcription (Speech to text (STT))
- Vosk
- Whisper
- OpenAI's Whisper model ported to C/C++ by ggerganov (Georgi Gerganov) | MIT License
Speech synthesis (text to speech (TTS))
Project | Credit | License |
---|---|---|
Bark | Suno AI | MIT License |
Coqui TTS | Coqui AI | MPL 2.0 |
SoftVC VITS Singing Voice Conversion Fork | voicepaw | Apache License 2.0 & MIT License |
TorToiSe | neonbjb (James Betker) | Apache License 2.0 |
Music
- Moseca separates music tracks into different stems (voice, drums, bass, guitar, piano, and others). This is useful for remixing, karaoke, and music studies. MIT
Uncategorized
- Virtual assistant: Mycroft
Permission is granted to copy, distribute and/or modify this document under the terms of the GNU Free Documentation License, Version 1.3 or any later version published by the Free Software Foundation; with no Invariant Sections, no Front-Cover Texts, and no Back-Cover Texts. A copy of the license is included in the page “GNU Free Documentation License”.
The copyright and license notices on this page only apply to the text on this page. Any software or copyright-licenses or other similar notices described in this text has its own copyright notice and license, which can usually be found in the distribution or license text itself.