43 Integrations with Azure AI Foundry Agent Service

View a list of Azure AI Foundry Agent Service integrations and software that integrates with Azure AI Foundry Agent Service below. Compare the best Azure AI Foundry Agent Service integrations as well as features, ratings, user reviews, and pricing of software that integrates with Azure AI Foundry Agent Service. Here are the current Azure AI Foundry Agent Service integrations in 2025:

  • 1
    Rayven

    Rayven

    Rayven

    Rayven's a full-stack, no/low-code platform that lets anyone build powerful apps, AI tools, and automations - fast. It's simple, scalable, and works with your existing technologies. Start building on our free-forever plan using AI prompts, drag-and-drop tools, templates, or code (when you want) to: - Build apps + automations in minutes using templates or from scratch - Connect systems, unify data, and integrate AI into your workflows - Modernise processes without replacing legacy tech or adding risk - Enable IT, Developers, DevOps, Ops, and Product teams to work from the same platform Rayven is designed for developers and non-technical users alike; giving you the all-in-one, low-cost tool to get from idea to outcome, fast. Start for free at rayven.io
    Starting Price: $0
  • 2
    Microsoft SharePoint
    Your mobile, intelligent intranet. Share and manage content, knowledge, and applications to empower teamwork, quickly find information, and seamlessly collaborate across the organization. SharePoint empowers teamwork with dynamic and productive team sites for every project team, department, and division. Share files, data, news, and resources. Customize your site to streamline your team’s work. Collaborate effortlessly and securely with team members inside and outside your organization, across PCs, Macs, and mobile devices. Build cohesion and inform your employees throughout your intranet. Drive organizational efficiency by sharing common resources and applications on home sites and portals. Tell your story with beautiful communication sites. And stay in the know with personalized, targeted news on the web and the SharePoint mobile apps. You’re just a click away from what you are looking for, with powerful search and intelligent ways to discover information.
    Leader badge
    Starting Price: $1.00/one-time/user
  • 3
    Microsoft Azure
    Microsoft's Azure is a cloud computing platform that allows for rapid and secure application development, testing and management. Azure. Invent with purpose. Turn ideas into solutions with more than 100 services to build, deploy, and manage applications—in the cloud, on-premises, and at the edge—using the tools and frameworks of your choice. Continuous innovation from Microsoft supports your development today, and your product visions for tomorrow. With a commitment to open source, and support for all languages and frameworks, build how you want, and deploy where you want to. On-premises, in the cloud, and at the edge—we’ll meet you where you are. Integrate and manage your environments with services designed for hybrid cloud. Get security from the ground up, backed by a team of experts, and proactive compliance trusted by enterprises, governments, and startups. The cloud you can trust, with the numbers to prove it.
  • 4
    OpenAI

    OpenAI

    OpenAI

    OpenAI’s mission is to ensure that artificial general intelligence (AGI)—by which we mean highly autonomous systems that outperform humans at most economically valuable work—benefits all of humanity. We will attempt to directly build safe and beneficial AGI, but will also consider our mission fulfilled if our work aids others to achieve this outcome. Apply our API to any language task — semantic search, summarization, sentiment analysis, content generation, translation, and more — with only a few examples or by specifying your task in English. One simple integration gives you access to our constantly-improving AI technology. Explore how you integrate with the API with these sample completions.
  • 5
    Bing

    Bing

    Microsoft

    Bing helps you turn information into action, making it faster and easier to go from searching to doing. Bing helps you turn information into action, making it faster and easier to go from searching to doing. As Bing continues to evolve, we’re focused on expanding the ways in which you can use Bing. For example, did you know that Bing powers InPrivate search with Microsoft Edge, quick searches in the Windows taskbar, work search scenarios with Microsoft 365, immersive gaming in Microsoft Flight Simulator, and much more? That’s why starting today, you will see a shift in product to Microsoft Bing, which reflects the continued integration of our search experiences across the Microsoft family. Beyond our commitment to expanding search scenarios, we also believe that you can and should get more value out of the searches you perform every day. That’s why we’re also excited to announce the expansion of Give with Bing, which helps you make a difference just by searching.
    Starting Price: Free
  • 6
    Mistral AI

    Mistral AI

    Mistral AI

    Mistral AI is a pioneering artificial intelligence startup specializing in open-source generative AI. The company offers a range of customizable, enterprise-grade AI solutions deployable across various platforms, including on-premises, cloud, edge, and devices. Flagship products include "Le Chat," a multilingual AI assistant designed to enhance productivity in both personal and professional contexts, and "La Plateforme," a developer platform that enables the creation and deployment of AI-powered applications. Committed to transparency and innovation, Mistral AI positions itself as a leading independent AI lab, contributing significantly to open-source AI and policy development.
    Starting Price: Free
  • 7
    Cohere

    Cohere

    Cohere AI

    Cohere is an enterprise AI platform that enables developers and businesses to build powerful language-based applications. Specializing in large language models (LLMs), Cohere provides solutions for text generation, summarization, and semantic search. Their model offerings include the Command family for high-performance language tasks and Aya Expanse for multilingual applications across 23 languages. Focused on security and customization, Cohere allows flexible deployment across major cloud providers, private cloud environments, or on-premises setups to meet diverse enterprise needs. The company collaborates with industry leaders like Oracle and Salesforce to integrate generative AI into business applications, improving automation and customer engagement. Additionally, Cohere For AI, their research lab, advances machine learning through open-source projects and a global research community.
    Starting Price: Free
  • 8
    GPT-4o

    GPT-4o

    OpenAI

    GPT-4o (“o” for “omni”) is a step towards much more natural human-computer interaction—it accepts as input any combination of text, audio, image, and video and generates any combination of text, audio, and image outputs. It can respond to audio inputs in as little as 232 milliseconds, with an average of 320 milliseconds, which is similar to human response time (opens in a new window) in a conversation. It matches GPT-4 Turbo performance on text in English and code, with significant improvement on text in non-English languages, while also being much faster and 50% cheaper in the API. GPT-4o is especially better at vision and audio understanding compared to existing models.
    Starting Price: $5.00 / 1M tokens
  • 9
    Azure AI Services
    Build cutting-edge, market-ready AI applications with out-of-the-box and customizable APIs and models. Quickly infuse generative AI into production workloads using studios, SDKs, and APIs. Gain a competitive edge by building AI apps powered by foundation models, including those from OpenAI, Meta, and Microsoft. Detect and mitigate harmful use with built-in responsible AI, enterprise-grade Azure security, and responsible AI tooling. Build your own copilot and generative AI applications with cutting-edge language and vision models. Retrieve the most relevant data using keyword, vector, and hybrid search. Monitor text and images to detect offensive or inappropriate content. Translate documents and text in real time across more than 100 languages.
  • 10
    OpenAI o1
    OpenAI o1 represents a new series of AI models designed by OpenAI, focusing on enhanced reasoning capabilities. These models, including o1-preview and o1-mini, are trained using a novel reinforcement learning approach to spend more time "thinking" through problems before providing answers. This approach allows o1 to excel in complex problem-solving tasks in areas like coding, mathematics, and science, outperforming previous models like GPT-4o in certain benchmarks. The o1 series aims to tackle challenges that require deeper thought processes, marking a significant step towards AI systems that can reason more like humans, although it's still in the preview stage with ongoing improvements and evaluations.
  • 11
    OpenAI o1-mini
    OpenAI o1-mini is a new, cost-effective AI model designed for enhanced reasoning, particularly excelling in STEM fields like mathematics and coding. It's part of the o1 series, which focuses on solving complex problems by spending more time "thinking" through solutions. Despite being smaller and 80% cheaper than its sibling, the o1-preview, o1-mini performs competitively in coding tasks and mathematical reasoning, making it an accessible option for developers and enterprises looking for efficient AI solutions.
  • 12
    Azure Blob Storage
    Massively scalable and secure object storage for cloud-native workloads, archives, data lakes, high-performance computing, and machine learning. Azure Blob Storage helps you create data lakes for your analytics needs, and provides storage to build powerful cloud-native and mobile apps. Optimize costs with tiered storage for your long-term data, and flexibly scale up for high-performance computing and machine learning workloads. Blob storage is built from the ground up to support the scale, security, and availability needs of mobile, web, and cloud-native application developers. Use it as a cornerstone for serverless architectures such as Azure Functions. Blob storage supports the most popular development frameworks, including Java, .NET, Python, and Node.js, and is the only cloud storage service that offers a premium, SSD-based object storage tier for low-latency and interactive scenarios.
    Starting Price: $0.00099
  • 13
    Microsoft Fabric
    Reshape how everyone accesses, manages, and acts on data and insights by connecting every data source and analytics service together—on a single, AI-powered platform. All your data. All your teams. All in one place. Establish an open and lake-centric hub that helps data engineers connect and curate data from different sources—eliminating sprawl and creating custom views for everyone. Accelerate analysis by developing AI models on a single foundation without data movement—reducing the time data scientists need to deliver value. Innovate faster by helping every person in your organization act on insights from within Microsoft 365 apps, such as Microsoft Excel and Microsoft Teams. Responsibly connect people and data using an open and scalable solution that gives data stewards additional control with built-in security, governance, and compliance.
    Starting Price: $156.334/month/2CU
  • 14
    Mistral 7B

    Mistral 7B

    Mistral AI

    Mistral 7B is a 7.3-billion-parameter language model that outperforms larger models like Llama 2 13B across various benchmarks. It employs Grouped-Query Attention (GQA) for faster inference and Sliding Window Attention (SWA) to efficiently handle longer sequences. Released under the Apache 2.0 license, Mistral 7B is accessible for deployment across diverse platforms, including local environments and major cloud services. Additionally, a fine-tuned version, Mistral 7B Instruct, demonstrates enhanced performance in instruction-following tasks, surpassing models like Llama 2 13B Chat.
    Starting Price: Free
  • 15
    Codestral Mamba
    As a tribute to Cleopatra, whose glorious destiny ended in tragic snake circumstances, we are proud to release Codestral Mamba, a Mamba2 language model specialized in code generation, available under an Apache 2.0 license. Codestral Mamba is another step in our effort to study and provide new architectures. It is available for free use, modification, and distribution, and we hope it will open new perspectives in architecture research. Mamba models offer the advantage of linear time inference and the theoretical ability to model sequences of infinite length. It allows users to engage with the model extensively with quick responses, irrespective of the input length. This efficiency is especially relevant for code productivity use cases, this is why we trained this model with advanced code and reasoning capabilities, enabling it to perform on par with SOTA transformer-based models.
    Starting Price: Free
  • 16
    Mistral NeMo

    Mistral NeMo

    Mistral AI

    Mistral NeMo, our new best small model. A state-of-the-art 12B model with 128k context length, and released under the Apache 2.0 license. Mistral NeMo is a 12B model built in collaboration with NVIDIA. Mistral NeMo offers a large context window of up to 128k tokens. Its reasoning, world knowledge, and coding accuracy are state-of-the-art in its size category. As it relies on standard architecture, Mistral NeMo is easy to use and a drop-in replacement in any system using Mistral 7B. We have released pre-trained base and instruction-tuned checkpoints under the Apache 2.0 license to promote adoption for researchers and enterprises. Mistral NeMo was trained with quantization awareness, enabling FP8 inference without any performance loss. The model is designed for global, multilingual applications. It is trained on function calling and has a large context window. Compared to Mistral 7B, it is much better at following precise instructions, reasoning, and handling multi-turn conversations.
    Starting Price: Free
  • 17
    Mixtral 8x22B

    Mixtral 8x22B

    Mistral AI

    Mixtral 8x22B is our latest open model. It sets a new standard for performance and efficiency within the AI community. It is a sparse Mixture-of-Experts (SMoE) model that uses only 39B active parameters out of 141B, offering unparalleled cost efficiency for its size. It is fluent in English, French, Italian, German, and Spanish. It has strong mathematics and coding capabilities. It is natively capable of function calling; along with the constrained output mode implemented on la Plateforme, this enables application development and tech stack modernization at scale. Its 64K tokens context window allows precise information recall from large documents. We build models that offer unmatched cost efficiency for their respective sizes, delivering the best performance-to-cost ratio within models provided by the community. Mixtral 8x22B is a natural continuation of our open model family. Its sparse activation patterns make it faster than any dense 70B model.
    Starting Price: Free
  • 18
    Mathstral

    Mathstral

    Mistral AI

    As a tribute to Archimedes, whose 2311th anniversary we’re celebrating this year, we are proud to release our first Mathstral model, a specific 7B model designed for math reasoning and scientific discovery. The model has a 32k context window published under the Apache 2.0 license. We’re contributing Mathstral to the science community to bolster efforts in advanced mathematical problems requiring complex, multi-step logical reasoning. The Mathstral release is part of our broader effort to support academic projects, it was produced in the context of our collaboration with Project Numina. Akin to Isaac Newton in his time, Mathstral stands on the shoulders of Mistral 7B and specializes in STEM subjects. It achieves state-of-the-art reasoning capacities in its size category across various industry-standard benchmarks. In particular, it achieves 56.6% on MATH and 63.47% on MMLU, with the following MMLU performance difference by subject between Mathstral 7B and Mistral 7B.
    Starting Price: Free
  • 19
    Ministral 3B

    Ministral 3B

    Mistral AI

    Mistral AI introduced two state-of-the-art models for on-device computing and edge use cases, named "les Ministraux": Ministral 3B and Ministral 8B. These models set a new frontier in knowledge, commonsense reasoning, function-calling, and efficiency in the sub-10B category. They can be used or tuned for various applications, from orchestrating agentic workflows to creating specialist task workers. Both models support up to 128k context length (currently 32k on vLLM), and Ministral 8B features a special interleaved sliding-window attention pattern for faster and memory-efficient inference. These models were built to provide a compute-efficient and low-latency solution for scenarios such as on-device translation, internet-less smart assistants, local analytics, and autonomous robotics. Used in conjunction with larger language models like Mistral Large, les Ministraux also serve as efficient intermediaries for function-calling in multi-step agentic workflows.
    Starting Price: Free
  • 20
    Ministral 8B

    Ministral 8B

    Mistral AI

    Mistral AI has introduced two advanced models for on-device computing and edge applications, named "les Ministraux": Ministral 3B and Ministral 8B. These models excel in knowledge, commonsense reasoning, function-calling, and efficiency within the sub-10B parameter range. They support up to 128k context length and are designed for various applications, including on-device translation, offline smart assistants, local analytics, and autonomous robotics. Ministral 8B features an interleaved sliding-window attention pattern for faster and more memory-efficient inference. Both models can function as intermediaries in multi-step agentic workflows, handling tasks like input parsing, task routing, and API calls based on user intent with low latency and cost. Benchmark evaluations indicate that les Ministraux consistently outperforms comparable models across multiple tasks. As of October 16, 2024, both models are available, with Ministral 8B priced at $0.1 per million tokens.
    Starting Price: Free
  • 21
    Mistral Small

    Mistral Small

    Mistral AI

    On September 17, 2024, Mistral AI announced several key updates to enhance the accessibility and performance of their AI offerings. They introduced a free tier on "La Plateforme," their serverless platform for tuning and deploying Mistral models as API endpoints, enabling developers to experiment and prototype at no cost. Additionally, Mistral AI reduced prices across their entire model lineup, with significant cuts such as a 50% reduction for Mistral Nemo and an 80% decrease for Mistral Small and Codestral, making advanced AI more cost-effective for users. The company also unveiled Mistral Small v24.09, a 22-billion-parameter model offering a balance between performance and efficiency, suitable for tasks like translation, summarization, and sentiment analysis. Furthermore, they made Pixtral 12B, a vision-capable model with image understanding capabilities, freely available on "Le Chat," allowing users to analyze and caption images without compromising text-based performance.
    Starting Price: Free
  • 22
    Azure OpenAI Service
    Apply advanced coding and language models to a variety of use cases. Leverage large-scale, generative AI models with deep understandings of language and code to enable new reasoning and comprehension capabilities for building cutting-edge applications. Apply these coding and language models to a variety of use cases, such as writing assistance, code generation, and reasoning over data. Detect and mitigate harmful use with built-in responsible AI and access enterprise-grade Azure security. Gain access to generative models that have been pretrained with trillions of words. Apply them to new scenarios including language, code, reasoning, inferencing, and comprehension. Customize generative models with labeled data for your specific scenario using a simple REST API. Fine-tune your model's hyperparameters to increase accuracy of outputs. Use the few-shot learning capability to provide the API with examples and achieve more relevant results.
    Starting Price: $0.0004 per 1000 tokens
  • 23
    Mixtral 8x7B

    Mixtral 8x7B

    Mistral AI

    Mixtral 8x7B is a high-quality sparse mixture of experts model (SMoE) with open weights. Licensed under Apache 2.0. Mixtral outperforms Llama 2 70B on most benchmarks with 6x faster inference. It is the strongest open-weight model with a permissive license and the best model overall regarding cost/performance trade-offs. In particular, it matches or outperforms GPT-3.5 on most standard benchmarks.
    Starting Price: Free
  • 24
    Llama 3
    We’ve integrated Llama 3 into Meta AI, our intelligent assistant, that expands the ways people can get things done, create and connect with Meta AI. You can see first-hand the performance of Llama 3 by using Meta AI for coding tasks and problem solving. Whether you're developing agents, or other AI-powered applications, Llama 3 in both 8B and 70B will offer the capabilities and flexibility you need to develop your ideas. With the release of Llama 3, we’ve updated the Responsible Use Guide (RUG) to provide the most comprehensive information on responsible development with LLMs. Our system-centric approach includes updates to our trust and safety tools with Llama Guard 2, optimized to support the newly announced taxonomy published by MLCommons expanding its coverage to a more comprehensive set of safety categories, code shield, and Cybersec Eval 2.
    Starting Price: Free
  • 25
    Codestral

    Codestral

    Mistral AI

    We introduce Codestral, our first-ever code model. Codestral is an open-weight generative AI model explicitly designed for code generation tasks. It helps developers write and interact with code through a shared instruction and completion API endpoint. As it masters code and English, it can be used to design advanced AI applications for software developers. Codestral is trained on a diverse dataset of 80+ programming languages, including the most popular ones, such as Python, Java, C, C++, JavaScript, and Bash. It also performs well on more specific ones like Swift and Fortran. This broad language base ensures Codestral can assist developers in various coding environments and projects.
    Starting Price: Free
  • 26
    Llama 3.1
    The open source AI model you can fine-tune, distill and deploy anywhere. Our latest instruction-tuned model is available in 8B, 70B and 405B versions. Using our open ecosystem, build faster with a selection of differentiated product offerings to support your use cases. Choose from real-time inference or batch inference services. Download model weights to further optimize cost per token. Adapt for your application, improve with synthetic data and deploy on-prem or in the cloud. Use Llama system components and extend the model using zero shot tool use and RAG to build agentic behaviors. Leverage 405B high quality data to improve specialized models for specific use cases.
    Starting Price: Free
  • 27
    Mistral Large

    Mistral Large

    Mistral AI

    Mistral Large is Mistral AI's flagship language model, designed for advanced text generation and complex multilingual reasoning tasks, including text comprehension, transformation, and code generation. It supports English, French, Spanish, German, and Italian, offering a nuanced understanding of grammar and cultural contexts. With a 32,000-token context window, it can accurately recall information from extensive documents. The model's precise instruction-following and native function-calling capabilities facilitate application development and tech stack modernization. Mistral Large is accessible through Mistral's platform, Azure AI Studio, and Azure Machine Learning, and can be self-deployed for sensitive use cases. Benchmark evaluations indicate that Mistral Large achieves strong results, making it the world's second-ranked model generally available through an API, next to GPT-4.
    Starting Price: Free
  • 28
    Azure AI Search
    Deliver high-quality responses with a vector database built for advanced retrieval augmented generation (RAG) and modern search. Focus on exponential growth with an enterprise-ready vector database that comes with security, compliance, and responsible AI practices built in. Build better applications with sophisticated retrieval strategies backed by decades of research and customer validation. Quickly deploy your generative AI app with seamless platform and data integrations for data sources, AI models, and frameworks. Automatically upload data from a wide range of supported Azure and third-party sources. Streamline vector data processing with built-in extraction, chunking, enrichment, and vectorization, all in one flow. Support for multivector, hybrid, multilingual, and metadata filtering. Move beyond vector-only search with keyword match scoring, reranking, geospatial search, and autocomplete.
    Starting Price: $0.11 per hour
  • 29
    Llama 3.2
    The open-source AI model you can fine-tune, distill and deploy anywhere is now available in more versions. Choose from 1B, 3B, 11B or 90B, or continue building with Llama 3.1. Llama 3.2 is a collection of large language models (LLMs) pretrained and fine-tuned in 1B and 3B sizes that are multilingual text only, and 11B and 90B sizes that take both text and image inputs and output text. Develop highly performative and efficient applications from our latest release. Use our 1B or 3B models for on device applications such as summarizing a discussion from your phone or calling on-device tools like calendar. Use our 11B or 90B models for image use cases such as transforming an existing image into something new or getting more information from an image of your surroundings.
    Starting Price: Free
  • 30
    Llama 3.3
    Llama 3.3 is the latest iteration in the Llama series of language models, developed to push the boundaries of AI-powered understanding and communication. With enhanced contextual reasoning, improved language generation, and advanced fine-tuning capabilities, Llama 3.3 is designed to deliver highly accurate, human-like responses across diverse applications. This version features a larger training dataset, refined algorithms for nuanced comprehension, and reduced biases compared to its predecessors. Llama 3.3 excels in tasks such as natural language understanding, creative writing, technical explanation, and multilingual communication, making it an indispensable tool for businesses, developers, and researchers. Its modular architecture allows for customizable deployment in specialized domains, ensuring versatility and performance at scale.
    Starting Price: Free
  • 31
    MacWhisper

    MacWhisper

    Gumroad

    ​MacWhisper enables users to quickly and easily transcribe audio files into text using OpenAI's Whisper technology. Users can record directly from their microphone or any input device on their Mac, or drag and drop audio files for high-quality transcription. It supports recording meetings from platforms like Zoom, Teams, Webex, Skype, Chime, and Discord, with all transcription processing done locally to ensure data privacy. Transcripts can be saved or exported in various formats, including .srt, .vtt, .csv, .docx, .pdf, markdown, and HTML. MacWhisper offers fast transcription speeds, supports over 100 languages, and provides features like search, audio playback synced to transcripts, filler word removal, and speaker addition. The Pro version includes additional functionalities such as batch transcription, YouTube video transcription, AI service integrations (e.g., OpenAI's ChatGPT, Anthropic's Claude), system-wide dictation, and translation of audio files into other languages.
    Starting Price: €59 one-time payment
  • 32
    Azure Logic Apps
    Built on a containerized runtime that increases scale and portability while automating business-critical workflows anywhere. Modernize your BizTalk Server applications by moving them to Logic Apps using the BizTalk migration tool. Connect logic apps to your virtual networks to seamlessly and securely integrate cloud-based and on-premises solutions. Containerize your workflows to deploy and run your applications anywhere—in the cloud, on premises, or the infrastructure of your choice. Apply CI/CD best practices to your workflows and take advantage of built-in tools for seamless and secure deployments. Deploy and run logic applications in Azure, any container, and on premises. Enable private endpoints, simplified virtual network access, and deployment slots. Develop, debug, and test on Windows, MacOS, and Linux using Visual Studio Code. Deploy multiple workflows to a single logic app, simplifying automated deployments and CI/CD pipelines.
  • 33
    Llama 2
    The next generation of our open source large language model. This release includes model weights and starting code for pretrained and fine-tuned Llama language models — ranging from 7B to 70B parameters. Llama 2 pretrained models are trained on 2 trillion tokens, and have double the context length than Llama 1. Its fine-tuned models have been trained on over 1 million human annotations. Llama 2 outperforms other open source language models on many external benchmarks, including reasoning, coding, proficiency, and knowledge tests. Llama 2 was pretrained on publicly available online data sources. The fine-tuned model, Llama-2-chat, leverages publicly available instruction datasets and over 1 million human annotations. We have a broad range of supporters around the world who believe in our open approach to today’s AI — companies that have given early feedback and are excited to build with Llama 2.
    Starting Price: Free
  • 34
    Pixtral Large

    Pixtral Large

    Mistral AI

    Pixtral Large is a 124-billion-parameter open-weight multimodal model developed by Mistral AI, building upon their Mistral Large 2 architecture. It integrates a 123-billion-parameter multimodal decoder with a 1-billion-parameter vision encoder, enabling advanced understanding of documents, charts, and natural images while maintaining leading text comprehension capabilities. With a context window of 128,000 tokens, Pixtral Large can process at least 30 high-resolution images simultaneously. The model has demonstrated state-of-the-art performance on benchmarks such as MathVista, DocVQA, and VQAv2, surpassing models like GPT-4o and Gemini-1.5 Pro. Pixtral Large is available under the Mistral Research License for research and educational use, and under the Mistral Commercial License for commercial applications.
    Starting Price: Free
  • 35
    Kong AI Gateway
    ​Kong AI Gateway is a semantic AI gateway designed to run and secure Large Language Model (LLM) traffic, enabling faster adoption of Generative AI (GenAI) through new semantic AI plugins for Kong Gateway. It allows users to easily integrate, secure, and monitor popular LLMs. The gateway enhances AI requests with semantic caching and security features, introducing advanced prompt engineering for compliance and governance. Developers can power existing AI applications written using SDKs or AI frameworks by simply changing one line of code, simplifying migration. Kong AI Gateway also offers no-code AI integrations, allowing users to transform, enrich, and augment API responses without writing code, using declarative configuration. It implements advanced prompt security by determining allowed behaviors and enables the creation of better prompts with AI templates compatible with the OpenAI interface.
  • 36
    Orq.ai

    Orq.ai

    Orq.ai

    Orq.ai is the #1 platform for software teams to operate agentic AI systems at scale. Optimize prompts, deploy use cases, and monitor performance, no blind spots, no vibe checks. Experiment with prompts and LLM configurations before moving to production. Evaluate agentic AI systems in offline environments. Roll out GenAI features to specific user groups with guardrails, data privacy safeguards, and advanced RAG pipelines. Visualize all events triggered by agents for fast debugging. Get granular control on cost, latency, and performance. Connect to your favorite AI models, or bring your own. Speed up your workflow with out-of-the-box components built for agentic AI systems. Manage core stages of the LLM app lifecycle in one central platform. Self-hosted or hybrid deployment with SOC 2 and GDPR compliance for enterprise security.
  • 37
    Azure Functions
    Develop more efficiently with Functions, an event-driven serverless compute platform that can also solve complex orchestration problems. Build and debug locally without additional setup, deploy and operate at scale in the cloud, and integrate services using triggers and bindings. End-to-end development experience with integrated tools and built-in DevOps capabilities. Integrated programming model to respond to events and seamlessly connect to other services. Implement a variety of functions and scenarios, such as web apps and APIs with .NET, Node.js, or Java; machine learning workflows with Python; and cloud automation with PowerShell. Get a complete serverless application development experience—from building and debugging locally to deploying and monitoring in the cloud.
  • 38
    Le Chat

    Le Chat

    Mistral AI

    Le Chat is a conversational entry point to interact with the various models from Mistral AI. It offers a pedagogical and fun way to explore Mistral AI’s technology. Le Chat can use Mistral Large or Mistral Small under the hood, or a prototype model called Mistral Next, designed to be brief and concise. We are hard at work to make our models as useful and as little opinionated as possible, although much remain to be improved! Thanks to a tunable system-level moderation mechanism, Le Chat warns you in a non-invasive way when you’re pushing the conversation in directions where the assistant may produce sensitive or controversial content.
    Starting Price: Free
  • 39
    Azure AI Foundry
    Azure AI Foundry is a unified application platform for your entire organization in the age of AI. Azure AI Foundry helps bridge the gap between cutting-edge AI technologies and practical business applications, empowering organizations to harness the full potential of AI efficiently and effectively. Azure AI Foundry is designed to empower your entire organization—developers, AI engineers, and IT professionals—to customize, host, run, and manage AI solutions with greater ease and confidence. This unified approach simplifies the development and management process, helping all stakeholders focus on driving innovation and achieving strategic goals. Azure AI Foundry Agent Service is a powerful component designed to facilitate the seamless operation of AI agents throughout the entire lifecycle—from development and deployment to production.
  • 40
    Orate

    Orate

    Orate

    Orate is an AI toolkit for speech that enables developers to create realistic, human-like speech and transcribe audio through a unified API compatible with leading AI providers such as OpenAI, ElevenLabs, and AssemblyAI. The platform offers text-to-speech functionality, allowing users to convert text into lifelike speech using a simple API that integrates seamlessly with various providers. For instance, by importing the 'speak' function from Orate and the desired provider, developers can generate speech from text prompts. Additionally, Orate provides speech-to-text capabilities, transforming spoken words into meaningful text with unparalleled accuracy, speed, and reliability. By importing the 'transcribe' function and the chosen provider, users can transcribe audio files into text. The toolkit also supports speech-to-speech transformations, enabling users to change the voice of their audio using a straightforward voice-to-voice API compatible with leading AI providers.
  • 41
    Aurascape

    Aurascape

    Aurascape

    ​Aurascape is an AI-native security platform designed to help businesses innovate securely in the age of AI. It provides comprehensive visibility into AI application interactions, safeguarding against data loss and AI-driven threats. Key features include monitoring AI activities across numerous applications, protecting sensitive data to ensure compliance, defending against zero-day threats, facilitating secure deployment of AI copilots, enforcing coding assistant guardrails, and automating AI security workflows. Aurascape's mission is to enable organizations to adopt AI technologies confidently while maintaining robust security measures. AI applications interact in fundamentally new ways. Communications are dynamic, real-time, and autonomous. Prevent new threats, protect data with unprecedented precision, and keep teams productive. Monitor unsanctioned app usage, risky authentication, and unsafe data sharing.
  • 42
    Phi-4-mini-flash-reasoning
    Phi-4-mini-flash-reasoning is a 3.8 billion‑parameter open model in Microsoft’s Phi family, purpose‑built for edge, mobile, and other resource‑constrained environments where compute, memory, and latency are tightly limited. It introduces the SambaY decoder‑hybrid‑decoder architecture with Gated Memory Units (GMUs) interleaved alongside Mamba state‑space and sliding‑window attention layers, delivering up to 10× higher throughput and a 2–3× reduction in latency compared to its predecessor without sacrificing advanced math and logic reasoning performance. Supporting a 64 K‑token context length and fine‑tuned on high‑quality synthetic data, it excels at long‑context retrieval, reasoning tasks, and real‑time inference, all deployable on a single GPU. Phi-4-mini-flash-reasoning is available today via Azure AI Foundry, NVIDIA API Catalog, and Hugging Face, enabling developers to build fast, scalable, logic‑intensive applications.
  • 43
    Llama

    Llama

    Meta

    Llama (Large Language Model Meta AI) is a state-of-the-art foundational large language model designed to help researchers advance their work in this subfield of AI. Smaller, more performant models such as Llama enable others in the research community who don’t have access to large amounts of infrastructure to study these models, further democratizing access in this important, fast-changing field. Training smaller foundation models like Llama is desirable in the large language model space because it requires far less computing power and resources to test new approaches, validate others’ work, and explore new use cases. Foundation models train on a large set of unlabeled data, which makes them ideal for fine-tuning for a variety of tasks. We are making Llama available at several sizes (7B, 13B, 33B, and 65B parameters) and also sharing a Llama model card that details how we built the model in keeping with our approach to Responsible AI practices.
  • Previous
  • You're on page 1
  • Next