Difference between revisions of "Free Software Directory:Artificial Intelligence Team"

From Free Software Directory
Jump to: navigation, search
m (added neat looking agnai project (multiple character chatbot))
(Image generation descriptions; removed StabilityAI listings: generally not a fan of the multiple updatings to Open RAIL model licenses when it is unclear how free-as-in-freedom they are (better safe than sorry).)
Line 82: Line 82:
 
| agnaistic  
 
| agnaistic  
 
| [https://github.com/agnaistic/agnai#AGPL-3.0-1-ov-file AGPLv3]
 
| [https://github.com/agnaistic/agnai#AGPL-3.0-1-ov-file AGPLv3]
| AI Agnostic (Multi-user and Multi-bot) Chat with Fictional Characters. Designed with scale in mind.
+
| AI agnostic (multi-user and multi-bot) chat with fictional characters
 
|-
 
|-
 
! [https://github.com/turboderp/exllamav2 ExLlamaV2]
 
! [https://github.com/turboderp/exllamav2 ExLlamaV2]
Line 117: Line 117:
 
| SillyTavern  
 
| SillyTavern  
 
| [https://github.com/SillyTavern/SillyTavern/blob/release/LICENSE AGPLv3]
 
| [https://github.com/SillyTavern/SillyTavern/blob/release/LICENSE AGPLv3]
| LLM Frontend for Power Users
+
| LLM frontend for power users
 
|-
 
|-
 
! [https://github.com/oobabooga/text-generation-webui Text Generation Web UI]  
 
! [https://github.com/oobabooga/text-generation-webui Text Generation Web UI]  
 
| oobabooga  
 
| oobabooga  
 
| [https://github.com/oobabooga/text-generation-webui/blob/main/LICENSE AGPLv3]
 
| [https://github.com/oobabooga/text-generation-webui/blob/main/LICENSE AGPLv3]
| A Gradio web UI for Large Language Models. Supports transformers, GPTQ, AWQ, EXL2, llama.cpp (GGUF), Llama models
+
| A Gradio web UI for Large Language Models. Supports transformers, GPTQ, AWQ, EXL2, llama.cpp (GGUF), LLaMA models
 
|-
 
|-
 
|}
 
|}
Line 166: Line 166:
 
| lm-sys  
 
| lm-sys  
 
| [https://github.com/lm-sys/FastChat/blob/main/LICENSE Apache-2.0]
 
| [https://github.com/lm-sys/FastChat/blob/main/LICENSE Apache-2.0]
| An open platform for training, serving, and evaluating large language models.
+
| An open platform for training, serving, and evaluating large language models
 
|-
 
|-
 
! [https://github.com/LAION-AI/Open-Assistant Open Assistant]  
 
! [https://github.com/LAION-AI/Open-Assistant Open Assistant]  
Line 176: Line 176:
 
| KillianLucas  
 
| KillianLucas  
 
| [https://github.com/KillianLucas/open-interpreter/blob/main/LICENSE AGPLv3]
 
| [https://github.com/KillianLucas/open-interpreter/blob/main/LICENSE AGPLv3]
| Lets LLMs run code (Python, Javascript, Shell, and more) locally.
+
| Lets LLMs run code (Python, Javascript, Shell, and more) locally
 
|-
 
|-
 
|}
 
|}
Line 191: Line 191:
 
== Images ==
 
== Images ==
 
=== Image generation ===
 
=== Image generation ===
==== Text to image ====
+
==== Text to image GUI ====
 
{| class="wikitable sortable"
 
{| class="wikitable sortable"
 
|-
 
|-
Line 197: Line 197:
 
! Credit  
 
! Credit  
 
! License
 
! License
 +
! Description
 
|-
 
|-
 
! [https://github.com/comfyanonymous/ComfyUI ComfyUI]  
 
! [https://github.com/comfyanonymous/ComfyUI ComfyUI]  
 
| comfyanonymous  
 
| comfyanonymous  
 
| [https://github.com/comfyanonymous/ComfyUI/blob/master/LICENSE GPLv3]
 
| [https://github.com/comfyanonymous/ComfyUI/blob/master/LICENSE GPLv3]
 +
| Modular Stable Diffusion GUI, API, and backend with a graph/nodes interface
 +
|-
 +
! [https://github.com/AUTOMATIC1111/stable-diffusion-webui Stable Diffusion WebUI]
 +
| AUTOMATIC1111
 +
| [https://github.com/AUTOMATIC1111/stable-diffusion-webui/blob/master/LICENSE.txt AGPLv3]
 +
| Stable Diffusion web UI
 +
|-
 +
|}
 +
==== Additional control libraries ====
 +
{| class="wikitable sortable"
 +
|-
 +
! Project
 +
! Credit
 +
! License
 +
! Description
 
|-
 
|-
 
! [https://github.com/lllyasviel/ControlNet ControlNet]  
 
! [https://github.com/lllyasviel/ControlNet ControlNet]  
 
| lllyasviel  
 
| lllyasviel  
 
| [https://github.com/lllyasviel/ControlNet/blob/main/LICENSE Apache-2.0]
 
| [https://github.com/lllyasviel/ControlNet/blob/main/LICENSE Apache-2.0]
 +
| Adding conditional control to text-to-image diffusion models
 +
|-
 +
! [https://github.com/huggingface/diffusers Diffusers]
 +
| huggingface
 +
| [https://github.com/huggingface/diffusers/blob/main/LICENSE Apache-2.0]
 +
| State-of-the-art diffusion models for image and audio generation in PyTorch
 +
|-
 +
|}
 +
==== Legacy ====
 +
{| class="wikitable sortable"
 +
|-
 +
! Project
 +
! Credit
 +
! License
 +
! Description
 
|-
 
|-
 
! [https://github.com/borisdayma/dalle-mini/ DALL-E Mini]  
 
! [https://github.com/borisdayma/dalle-mini/ DALL-E Mini]  
 
| borisdayma (Boris Dayma)  
 
| borisdayma (Boris Dayma)  
 
| [https://github.com/borisdayma/dalle-mini/blob/main/LICENSE Apache 2.0]
 
| [https://github.com/borisdayma/dalle-mini/blob/main/LICENSE Apache 2.0]
|-
+
| Generate images from a text prompt
! [https://github.com/huggingface/diffusers Diffusers]
 
| huggingface
 
| [https://github.com/huggingface/diffusers/blob/main/LICENSE Apache-2.0]
 
 
|-
 
|-
 
! [https://github.com/anishathalye/neural-style neural-style]  
 
! [https://github.com/anishathalye/neural-style neural-style]  
 
| anishathalye  
 
| anishathalye  
 
| [https://github.com/anishathalye/neural-style/blob/master/LICENSE.txt GPLv3]
 
| [https://github.com/anishathalye/neural-style/blob/master/LICENSE.txt GPLv3]
|-
+
| An implementation of neural style in TensorFlow
! [https://github.com/Stability-AI/generative-models SDXL - generative-models]
 
| Stability-AI
 
| [https://github.com/Stability-AI/generative-models/blob/main/LICENSE-CODE MIT]
 
|-
 
! [https://github.com/AUTOMATIC1111/stable-diffusion-webui Stable Diffusion WebUI]
 
| AUTOMATIC1111
 
| [https://github.com/AUTOMATIC1111/stable-diffusion-webui/blob/master/LICENSE.txt AGPLv3]
 
 
|-
 
|-
 
|}
 
|}
  
====Videos====
+
===Videos===
 
Image generation techniques create pictures from noise estimations. This noise shows up as artifacts and hampers temporal stability for objects. These projects tackle that issue.
 
Image generation techniques create pictures from noise estimations. This noise shows up as artifacts and hampers temporal stability for objects. These projects tackle that issue.
 
{| class="wikitable sortable"
 
{| class="wikitable sortable"
Line 235: Line 256:
 
! Credit  
 
! Credit  
 
! License
 
! License
 +
! Description
 
|-
 
|-
 
! [https://github.com/google-research/frame-interpolation FILM: Frame Interpolation for Large Motion]  
 
! [https://github.com/google-research/frame-interpolation FILM: Frame Interpolation for Large Motion]  
 
| Google Research
 
| Google Research
 
| [https://github.com/google-research/frame-interpolation#Apache-2.0-1-ov-file Apache 2.0]
 
| [https://github.com/google-research/frame-interpolation#Apache-2.0-1-ov-file Apache 2.0]
 +
| A unified single-network approach to frame interpolation that doesn't use additional pre-trained networks
 
|-
 
|-
! [https://github.com/magic-research/magic-animate MagicAnimate: Temporally Consistent Human Image Animation using Diffusion Model]
+
! [https://github.com/magic-research/magic-animate MagicAnimate]
 
| MagIC Research
 
| MagIC Research
 
| [https://github.com/magic-research/magic-animate#BSD-3-Clause-1-ov-file  BSD-3-Clause license]
 
| [https://github.com/magic-research/magic-animate#BSD-3-Clause-1-ov-file  BSD-3-Clause license]
 +
| Temporally consistent human image animation using a diffusion model
 
|-
 
|-
! [https://github.com/TencentARC/MotionCtrl MotionCtrl: A Unified and Flexible Motion Controller for Video Generation]
+
! [https://github.com/TencentARC/MotionCtrl MotionCtrl]
 
| ARC Lab, Tencent PCG
 
| ARC Lab, Tencent PCG
 
| [https://github.com/TencentARC/MotionCtrl#Apache-2.0-1-ov-file Apache 2.0]
 
| [https://github.com/TencentARC/MotionCtrl#Apache-2.0-1-ov-file Apache 2.0]
|-
+
| A unified and flexible motion controller for video generation
! [https://stability.ai/news/stable-video-diffusion-open-ai-video-model Stable Video Diffusion]
 
| Stability AI
 
| [https://github.com/Stability-AI/generative-models#MIT-1-ov-file MIT License]
 
 
|-
 
|-
 
! [https://github.com/CiaraStrawberry/TemporalKit TemporalKit]  
 
! [https://github.com/CiaraStrawberry/TemporalKit TemporalKit]  
 
| CiaraStrawberry (Ciara Rowles)
 
| CiaraStrawberry (Ciara Rowles)
 
| [https://github.com/CiaraStrawberry/TemporalKit?tab=GPL-3.0-1-ov-file#readme GPLv3]
 
| [https://github.com/CiaraStrawberry/TemporalKit?tab=GPL-3.0-1-ov-file#readme GPLv3]
 +
| An all in one solution for adding temporal stability to a Stable Diffusion render via an automatic1111 extension
 
|-
 
|-
 
! [https://github.com/camenduru/text-to-video-synthesis-colab Text To Video Synthesis Colab]  
 
! [https://github.com/camenduru/text-to-video-synthesis-colab Text To Video Synthesis Colab]  
 
| camenduru  
 
| camenduru  
 
| [https://github.com/camenduru/text-to-video-synthesis-colab/blob/main/LICENSE The Unlicense]
 
| [https://github.com/camenduru/text-to-video-synthesis-colab/blob/main/LICENSE The Unlicense]
 +
| A text-to-video synthesis model that evolves from a text-to-image synthesis model
 
|-
 
|-
 
! [https://github.com/yoyo-nb/Thin-Plate-Spline-Motion-Model Thin-Plate Spline Motion Model for Image Animation]  
 
! [https://github.com/yoyo-nb/Thin-Plate-Spline-Motion-Model Thin-Plate Spline Motion Model for Image Animation]  
 
| yoyo-nb  
 
| yoyo-nb  
 
| [https://github.com/yoyo-nb/Thin-Plate-Spline-Motion-Model/blob/main/LICENSE MIT]
 
| [https://github.com/yoyo-nb/Thin-Plate-Spline-Motion-Model/blob/main/LICENSE MIT]
 +
| Animates a static object in a source image according to a driving video
 
|-
 
|-
 
|}
 
|}

Revision as of 01:55, 2 March 2024

Free Software Foundation-Free Software Directory-Artificial Intelligence Project Team.png

The Artificial Intelligence Project Team gathers free software resources regarding machine learning / artificial intelligence.

Group info User info
User Role Reference Real name libera.chat nick Time zone Title
David_Hedlund Coordinator David Hedlund David_Hedlund Europe/Stockholm
GrahamxReed Collaborator Graham Reed Graham_Reed America/New_York
Mertgor Observer Mert Gör hwpplayer1 Europe/Istanbul
Mmcmahon Team captain Michael McMahon thomzane America/New_York FSF Systems Administrator

Truthfulness

Transformers

Text

Grammar

Translation

Text generation

Front ends

Project Credit License Description
Agnai agnaistic AGPLv3 AI agnostic (multi-user and multi-bot) chat with fictional characters
ExLlamaV2 turboderp MIT License A fast inference library for running LLMs locally on modern consumer-class GPUs
ExLlamaV2 WebUI turboderp MIT License
GPT4All Nomic AI MIT Run open-source LLMs anywhere
KoboldAI KoboldAI AGPLv3 A browser-based front-end for AI-assisted writing with multiple local & remote AI models
koboldcpp LostRuins AGPLv3 A simple one-file way to run various GGML and GGUF models with KoboldAI's UI
Serge serge-chat MIT A web interface for chatting with Alpaca through llama.cpp. Fully dockerized, with an easy to use API.
SillyTavern SillyTavern AGPLv3 LLM frontend for power users
Text Generation Web UI oobabooga AGPLv3 A Gradio web UI for Large Language Models. Supports transformers, GPTQ, AWQ, EXL2, llama.cpp (GGUF), LLaMA models

Models

Project Credit License Description
llama.cpp ggerganov (Georgi Gerganov) MIT Port of Facebook's LLaMA model in C/C++
Mistral Mistral AI Apache 2.0 Mistral 7B significantly outperforms Llama 2 13B on all metrics, and is on par with Llama 34B.
Mixtral Mistral AI Apache 2.0 Mixtral outperforms Llama 2 70B on most benchmarks with 6x faster inference.

Other

Project Credit License Description
GPT-NeoX EleutherAI Apache 2.0 EleutherAI's library for training large-scale language models on GPUs
FastChat lm-sys Apache-2.0 An open platform for training, serving, and evaluating large language models
Open Assistant LAION-AI Apache 2.0 A chat-based assistant that understands tasks, can interact with third-party systems, and retrieve information dynamically to do so
Open Interpreter KillianLucas AGPLv3 Lets LLMs run code (Python, Javascript, Shell, and more) locally

Honorary mention: LLaMA 1 had an AGPLv3 license.

Code generation

This concept is controversial. See the FSF's other writing on this topic.

Images

Image generation

Text to image GUI

Project Credit License Description
ComfyUI comfyanonymous GPLv3 Modular Stable Diffusion GUI, API, and backend with a graph/nodes interface
Stable Diffusion WebUI AUTOMATIC1111 AGPLv3 Stable Diffusion web UI

Additional control libraries

Project Credit License Description
ControlNet lllyasviel Apache-2.0 Adding conditional control to text-to-image diffusion models
Diffusers huggingface Apache-2.0 State-of-the-art diffusion models for image and audio generation in PyTorch

Legacy

Project Credit License Description
DALL-E Mini borisdayma (Boris Dayma) Apache 2.0 Generate images from a text prompt
neural-style anishathalye GPLv3 An implementation of neural style in TensorFlow

Videos

Image generation techniques create pictures from noise estimations. This noise shows up as artifacts and hampers temporal stability for objects. These projects tackle that issue.

Project Credit License Description
FILM: Frame Interpolation for Large Motion Google Research Apache 2.0 A unified single-network approach to frame interpolation that doesn't use additional pre-trained networks
MagicAnimate MagIC Research BSD-3-Clause license Temporally consistent human image animation using a diffusion model
MotionCtrl ARC Lab, Tencent PCG Apache 2.0 A unified and flexible motion controller for video generation
TemporalKit CiaraStrawberry (Ciara Rowles) GPLv3 An all in one solution for adding temporal stability to a Stable Diffusion render via an automatic1111 extension
Text To Video Synthesis Colab camenduru The Unlicense A text-to-video synthesis model that evolves from a text-to-image synthesis model
Thin-Plate Spline Motion Model for Image Animation yoyo-nb MIT Animates a static object in a source image according to a driving video


Image captioning

3D modeling

Audio

Natural language processing (NLP)

Transcription (Speech to text (STT))

Speech synthesis (text to speech (TTS))

Project Credit License
Bark Suno AI MIT License
Coqui TTS Coqui AI MPL 2.0
SoftVC VITS Singing Voice Conversion Fork voicepaw Apache License 2.0 & MIT License
TorToiSe neonbjb (James Betker) Apache License 2.0

Music

  • Moseca separates music tracks into different stems (voice, drums, bass, guitar, piano, and others). This is useful for remixing, karaoke, and music studies. MIT

Uncategorized



Permission is granted to copy, distribute and/or modify this document under the terms of the GNU Free Documentation License, Version 1.3 or any later version published by the Free Software Foundation; with no Invariant Sections, no Front-Cover Texts, and no Back-Cover Texts. A copy of the license is included in the page “GNU Free Documentation License”.

The copyright and license notices on this page only apply to the text on this page. Any software or copyright-licenses or other similar notices described in this text has its own copyright notice and license, which can usually be found in the distribution or license text itself.