AI/ML Products: Open Source Models, Frameworks & Tools

Explore open source artificial intelligence and machine learning products. From LLMs to AI infrastructure, enjoy AI innovation without vendor constraints.

AI/ML Solutions for Open Innovation

Discover tools and frameworks to perfect your AI journey. From cutting-edge models to robust infrastructure, we've got everything you need to accelerate innovation.

Don’t see the open source product you need? Contact us—we likely have it.

Large LLMs

🚀 Optimized for advanced use cases requiring high-scale performance.

Developed by Meta, this large model is competitive with proprietary systems, known for its strong performance across various tasks.

From Mistral AI, notable for its efficiency and competitive performance.

Utilizes a hybrid Transformer-Mamba mixture of expert architecture, offering high throughput and low memory usage with an effective context length of 256K tokens.

Features a mixture of expert design, activating 21 billion parameters per token, optimized for efficient processing.

A model focusing on following user directives accurately, specializing in few-turn interactions.

Employs a sparse mixture of expert architecture, engaging 39 billion active parameters per token, balancing performance and efficiency.

Designed for general-purpose tasks.

Leading space under the cc-by-nc-4.0 license, this model is tailored to long context and multi-step agentic RAG with tool use.

From NVIDIA, a large general-purpose model with extensive alignment.

Mid-Sized LLMs

🛠️ Balanced for efficiency and performance in smaller environments.

New in Meta’s Llama series, offering a balance between performance and resource requirements.

Designed for efficient processing with strong performance in various applications.

A model optimized for multilingual support and diverse task performance.

Focuses on conversational AI applications, providing coherent and contextually relevant responses.

Utilizes a sparse mixture of experts architecture with 8 experts of 7B parameters each, engaging 12.9B active parameters per token for efficient processing.

Code/Logic LLMs

💻 Models optimized for code generation and logical reasoning.

A model specialized in code generation and understanding, supporting multiple programming languages.

Qwen with Questions, a preview model focused on internal-dialog-like logical reasoning.

A smaller model designed for code generation across various programming languages.

Mistral AI’s model focused on code generation, surpassing larger models like Llama3 70B in the HumanEval FIM benchmark.

Variants of Meta’s Llama models fine-tuned for code generation and understanding.

A model optimized for code generation and conversational AI, facilitating interactive coding assistance.

Vision Language Models (VLMs)

🖼️ Bridging visual and textual understanding.

A model combining the SigLIP-So400m vision encoder with the Gemma-2B language model, designed for versatile vision-language tasks.

Developed by CogAI, focusing on integrating visual and textual data for comprehensive understanding.

An enhanced version of Qwen-VL, incorporating a Vision Transformer (ViT) for seamless image and video input processing, achieving superior performance across various tasks.

A series of multimodal models available in 7B and 72B parameter sizes, designed for advanced vision-language understanding.

A multimodal model trained to understand both natural images and documents, achieving leading performance on various multimodal benchmarks without compromising on text capabilities.

A 4.2-billion-parameter multimodal model by Microsoft, optimized for efficiency, capable of running on devices with limited computational resources, including modern smartphones.

Smaller LLMs

🔍 Lightweight models for rapid deployment.

The smallest of Meta’s Llama 3.1 series, offering competitive performance in a lightweight package, suitable for various applications.

From Google’s Gemma 2 series, designed to outperform models like Llama-3 8B in several benchmarks, offering efficiency and competitiveness.
A model balancing performance and cost efficiency, suitable for tasks like text summarization and code completion.
A hybrid Mamba2-Transformers architecture that offers improved memory efficiency and faster inference latency compared to transformer-only models, while outperforming similarly-sized models in benchmarks.

A compact hybrid Mamba2-Transformers model optimized for efficiency, offering reduced memory overhead, faster inference, and superior performance across benchmarks compared to other models in its size class.

AI/ML Infrastructure

Power your AI models with the right tools for performance and scalability.

JAX:

High-performance machine learning with automatic differentiation.

PyTorch:

A leading deep learning framework.

Ray:

Distributed computing made easy.

Dask:

Flexible parallel computing.

Hugging Face:

Democratizing AI development (some models require purchased licensing).

Lightning:

Streamlining deep learning research (some models require purchased licensing).

Data for AI

Organize and process your data with powerful tools and databases.

Data Processing Tools

Kafka:

Real-time data streaming.

Spark:

Unified analytics for large-scale data.

Flink:

Stream processing at its finest.

Airflow:

Workflow automation made simple.

Dask:

Parallel computing for large datasets.

Presto:

Interactive querying at scale.

ElasticSearch:

Advanced search and analytics.

Iceberg:

Modern table format for data lakes.

Vector Databases

Efficiently store and query embeddings.

Pinecone, Weaviate, Milvus, and more.

AI Ops

Deploy, manage, and monitor your AI applications with confidence.

Nebari:

Your foundation for scalable AI operations.

Conda:

Streamlining package management and environment creation for reproducible workflows.

VLLM & Triton:

Simplifying deployment and inference at scale.

Seamless AI Stack Integration

We make all the pieces work together. Whether you’re building pipelines, deploying models, or integrating AI solutions, our team ensures smooth implementation for your stack.

Let’s Talk

Looking for something specific or need guidance on choosing the right tools? We want to hear from you!