Found 16,907 repositories(showing 30)
huggingface
The largest collection of PyTorch image encoders / backbones. Including train, eval, inference, export scripts, and pretrained weights -- ResNet, ResNeXT, EfficientNet, NFNet, Vision Transformer (ViT), MobileNetV4, MobileNet-V3 & V2, RegNet, DPN, CSPNet, Swin Transformer, MaxViT, CoAtNet, ConvNeXt, and more
langfuse
๐ชข Open source LLM engineering platform: LLM Observability, metrics, evals, prompt management, playground, datasets. Integrates with OpenTelemetry, Langchain, OpenAI SDK, LiteLLM, and more. ๐YC W23
mastra-ai
From the team behind Gatsby, Mastra is a framework for building AI-powered applications and agents with a modern TypeScript stack.
openai
Evals is a framework for evaluating LLMs and LLM systems, and an open-source registry of benchmarks.
ConardLi
A powerful tool for creating datasets for LLM fine-tuning ใRAG and Eval
Arize-ai
AI Observability & Evaluation
allenai
Modeling, training, eval, and inference code for OLMo
AgentOps-AI
Python SDK for AI agent monitoring, LLM cost tracking, benchmarking, and more. Integrates with most LLMs and agent frameworks including CrewAI, Agno, OpenAI Agents SDK, Langchain, Autogen, AG2, and CamelAI
Kiln-AI
Build, Evaluate, and Optimize AI Systems. Includes evals, RAG, agents, fine-tuning, synthetic data generation, dataset management, MCP, and more.
openai
No description available
pydantic
AI observability platform for production LLM and agent systems.
PrimeIntellect-ai
Our library for RL environments + evals
MoonHighway
The code samples for Learning React by Alex Banks and Eve Porcello, published by O'Reilly Media
truera
Evaluation and Tracking for LLM Experiments and AI Agents
lmnr-ai
Laminar - open-source observability platform purpose-built for AI agents. YC S24.
modelscope
A streamlined and customizable framework for efficient large model (LLM, VLM, AIGC) evaluation and performance benchmarking.
hegdepavankumar
Free Images for EVE-NG and GNS3 containing routers, switches,Firewalls and other appliances, including Cisco, Fortigate, Palo Alto, Sophos and more. Master the art of networking and improve your skills!, our repository provides a one-stop solution for a comprehensive hands-on experience.
cosmos72
Interactive Go interpreter and debugger with REPL, Eval, generics and Lisp-like macros
hkust-nlp
Official github repo for C-Eval, a Chinese evaluation suite for foundation models [NeurIPS 2023]
pyfa-org
Python fitting assistant, cross-platform fitting tool for EVE Online
dreamzero0
Code to pretrain, fine-tune, and evaluate DreamZero and run sim & real-world evals
GitHamza0206
OpenSource Production ready Customer service with built in Evals and monitoring
mattpocock
Evaluate your LLM-powered apps with TypeScript
harbor-framework
Harbor is a framework for running agent evaluations and creating and using RL environments.
albertlatacz
Read Eval Print Loop for Java
openai
OpenAI Frontier Evals
superlinear-ai
๐ฅค RAGLite is a Python toolkit for Retrieval-Augmented Generation (RAG) with DuckDB or PostgreSQL
facebookarchive
A read-eval-print-loop for php
hamelsmu
Skills for AI Evals to compliment the course: AI Evals For Engineers & PMs
JudgmentLabs
The open source post-building layer for agents. Our environment data and evals power agent post-training (RL, SFT) and monitoring.