AI-Powered LLM Platform
for Enterprise Solutions

TraceMyPods delivers powerful AI capabilities through a secure, scalable platform with multiple LLM models and image generation capabilities.

Platform Features

TraceMyPods combines powerful AI capabilities with enterprise-grade infrastructure

πŸ“¦

Modular Microservices

Fine-grained APIs like admin, order, token, ask, and deliver services make the platform highly modular and maintainable.

🧩

Embeddings & Vector Search

Advanced search powered by Qdrant and custom embeddings from embedding-api for real-time semantic search and AI memory.

πŸ“¨

Email Sending with Resend

Built-in SMTP support for OTP verification

πŸ”

Secure Token Authentication

Generate secure tokens for API access with Redis-backed authentication and 1-hour expiry for enhanced security.

🧠

Multiple AI Models

Access a variety of LLM models from TinyLlama to powerful Mistral and CodeLlama for different use cases and requirements.

πŸ–ΌοΈ

Image Generation

Create AI-generated images from text descriptions with our public API feature, currently in beta.

⚑

High Performance

Optimized infrastructure with GPU acceleration for AI models and efficient request routing.

πŸ”§

Customizable

Easily extendable and customizable to fit your specific needs with a modular architecture.

πŸ“Š

Analytics Dashboard

Comprehensive analytics dashboard for monitoring usage, performance, and model interactions.

Available Models

Choose from our selection of powerful AI models to suit your specific needs

πŸ’‘
Smoll
~500 Mb 1 GB RAM

Free Lightweight model perfect for chat bot with minimal resource requirements.

Free
🌐
Gemma 2B
2.6 GB 6–8 GB RAM

Google’s open-weight chat-optimized model suitable for small to medium workloads.

Open Source
πŸ¦…
Falcon-RW 1B
~1.3 GB 4 GB RAM

Small version of the Falcon family, ideal for offline summarization and QA tasks.

Apache 2.0
πŸ§‘β€πŸ’»
Replit Code 3B
3.3 GB 8–10 GB RAM

Fine-tuned for code generation and completions. Great for coding copilots.

Open Source
🀏
TinyLlama
~1.1 GB 4 GB RAM

Lightweight model perfect for simple Q&A and chat applications with minimal resource requirements.

$15/month
🧠
Mistral-7B
~4.2 GB 8-16 GB RAM

Powerful general-purpose model with excellent reasoning capabilities and broad knowledge.

$15/month
πŸ’»
CodeLlama
4.5-10 GB 16-24 GB RAM

Specialized for code generation and understanding across multiple programming languages.

$10/month
πŸ¦™
LLaMA 2
4.5-40 GB 16-80 GB RAM

Versatile but resource-heavy model with state-of-the-art performance across various tasks.

$20/month
Ο†
Phi-2
~1.7 GB 6-8 GB RAM

Efficient and compact model with excellent reasoning capabilities for its size.

$12/month

πŸ€–
TraceMyPods AI Assistant

Token: Not Generated
Waiting for your query...

πŸ’‘ Premium models available for enhanced capabilities:

#mistral #codellama #llama2 #phi

↑