7 Integrations with Portkey

View a list of Portkey integrations and software that integrates with Portkey below. Compare the best Portkey integrations as well as features, ratings, user reviews, and pricing of software that integrates with Portkey. Here are the current Portkey integrations in 2024:

  • 1
    OpenAI

    OpenAI

    OpenAI

    OpenAI’s mission is to ensure that artificial general intelligence (AGI)—by which we mean highly autonomous systems that outperform humans at most economically valuable work—benefits all of humanity. We will attempt to directly build safe and beneficial AGI, but will also consider our mission fulfilled if our work aids others to achieve this outcome. Apply our API to any language task — semantic search, summarization, sentiment analysis, content generation, translation, and more — with only a few examples or by specifying your task in English. One simple integration gives you access to our constantly-improving AI technology. Explore how you integrate with the API with these sample completions.
  • 2
    Claude

    Claude

    Anthropic

    Claude is an artificial intelligence large language model that can process and generate human-like text. Anthropic is an AI safety and research company that’s working to build reliable, interpretable, and steerable AI systems. Large, general systems of today can have significant benefits, but can also be unpredictable, unreliable, and opaque: our goal is to make progress on these issues. For now, we’re primarily focused on research towards these goals; down the road, we foresee many opportunities for our work to create value commercially and for public benefit.
    Starting Price: Free
  • 3
    NGINX

    NGINX

    F5

    NGINX Open Source: The open source web server that powers more than 400 million websites. NGINX Plus is a software load balancer, web server, and content cache built on top of open source NGINX. Use NGINX Plus instead of your hardware load balancer and get the freedom to innovate without being constrained by infrastructure. Save more than 80% compared to hardware ADCs, without sacrificing performance or functionality. Deploy anywhere: public cloud, private cloud, bare metal, virtual machines, and containers. Save time by performing common tasks through the built‑in NGINX Plus API. From NetOps to DevOps, modern app teams need a self‑service, API‑driven platform that integrates easily into CI/CD workflows to accelerate app deployment – whether your app has a hybrid or microservices architecture – and makes app lifecycle management easier.
  • 4
    Azure OpenAI Service
    Apply advanced coding and language models to a variety of use cases. Leverage large-scale, generative AI models with deep understandings of language and code to enable new reasoning and comprehension capabilities for building cutting-edge applications. Apply these coding and language models to a variety of use cases, such as writing assistance, code generation, and reasoning over data. Detect and mitigate harmful use with built-in responsible AI and access enterprise-grade Azure security. Gain access to generative models that have been pretrained with trillions of words. Apply them to new scenarios including language, code, reasoning, inferencing, and comprehension. Customize generative models with labeled data for your specific scenario using a simple REST API. Fine-tune your model's hyperparameters to increase accuracy of outputs. Use the few-shot learning capability to provide the API with examples and achieve more relevant results.
    Starting Price: $0.0004 per 1000 tokens
  • 5
    Mixtral 8x7B

    Mixtral 8x7B

    Mistral AI

    Mixtral 8x7B is a high-quality sparse mixture of experts model (SMoE) with open weights. Licensed under Apache 2.0. Mixtral outperforms Llama 2 70B on most benchmarks with 6x faster inference. It is the strongest open-weight model with a permissive license and the best model overall regarding cost/performance trade-offs. In particular, it matches or outperforms GPT-3.5 on most standard benchmarks.
    Starting Price: Free
  • 6
    Amazon Bedrock
    The easiest way to build and scale generative AI applications with foundation models (FMs) Amazon Bedrock provides you the flexibility to choose from a wide range of FMs built by leading AI startups and Amazon so you can find the model that is best suited for what you are trying to get done. With Bedrock’s serverless experience, you can get started quickly, privately customize FMs with your own data, and easily integrate and deploy them into your applications using the AWS tools and capabilities you are familiar with (including integrations with Amazon SageMaker ML features like Experiments to test different models and Pipelines to manage your FMs at scale) without having to manage any infrastructure. Choose FMs from AI21 Labs, Anthropic, Stability AI, and Amazon to find the right FM for your use case.
  • 7
    Groq

    Groq

    Groq

    Groq is on a mission to set the standard for GenAI inference speed, helping real-time AI applications come to life today. An LPU inference engine, with LPU standing for Language Processing Unit, is a new type of end-to-end processing unit system that provides the fastest inference for computationally intensive applications with a sequential component, such as AI language applications (LLMs). The LPU is designed to overcome the two LLM bottlenecks, compute density and memory bandwidth. An LPU has greater computing capacity than a GPU and CPU in regards to LLMs. This reduces the amount of time per word calculated, allowing sequences of text to be generated much faster. Additionally, eliminating external memory bottlenecks enables the LPU inference engine to deliver orders of magnitude better performance on LLMs compared to GPUs. Groq supports standard machine learning frameworks such as PyTorch, TensorFlow, and ONNX for inference.
  • Previous
  • You're on page 1
  • Next