Gcore Announces Inference at the Edge – Bringing AI Applications Closer to End Users for Seamless Real-Time Performance

Applications of AI


New AI solution enables fast, secure and cost-effective deployment of pre-trained machine learning models at the edge, globally

CD 3590 1CD 3590 1
GCORE Orange 001GCORE Orange 001
GCORE Orange 001GCORE Orange 001

LUXEMBOURG–(BUSINESS WIRE)–#AI—Gcore, a global provider of edge AI, cloud, network and security solutions, today announced the release of Gcore Inference at the Edge, a groundbreaking solution that delivers an ultra-low latency experience for AI applications. The innovative solution enables distributed deployment of pre-trained machine learning (ML) models to edge inference nodes, ensuring seamless, real-time inference.

Gcore Inference at the Edge provides cost-effective, scalable and secure deployment of AI models for enterprises across industries including automotive, manufacturing, retail and technology. Use cases such as generative AI, object recognition, real-time behavioral analysis, virtual assistants and production monitoring can now be rapidly realized at global scale.

Gcore Inference at the Edge runs on Gcore's extensive global network of over 180 edge nodes, all interconnected by Gcore's advanced low-latency smart routing technology. Each high-performance node is located at the edge of the Gcore network, strategically placing servers close to end users. Inference at the Edge runs on NVIDIA L40S GPUs, the market-leading chips designed specifically for AI inference. When a user submits a request, the edge node determines the route to the nearest available inference area with the lowest latency, delivering typical response times of less than 30 milliseconds.

The new solution supports a wide range of foundational ML models as well as custom models. Open-source foundational models available on the Gcore ML model hub include LLaMA Pro 8B, Mistral 7B, and Stable-Diffusion XL. Models can be selected and trained for any use case and then distributed globally to Gcore Inference on edge nodes. This addresses a major challenge faced by development teams, where AI models typically run on the same servers used for training, resulting in reduced performance.

– Message from our partners –

The benefits of Gcore inference at the edge include:

  • Cost-effective deployment: Our flexible pricing structure ensures you pay only for the resources you use.
  • Built-in DDoS protection: ML endpoints are automatically protected from DDoS attacks through Gcore's infrastructure.
  • Superior data privacy and security: The solution comes with built-in compliance with GDPR, PCI DSS, and ISO/IEC 27001 standards.
  • Autoscaling the model: Autoscaling is available to handle sudden load spikes, ensuring your models are always ready for peak demand or unexpected surges.
  • Unlimited Object Storage: Scalable S3-compatible cloud storage that can grow as your model needs change.

Gcore CEO Andre Reitenbach Comment: “Gcore Inference at the Edge enables our customers to focus on training their machine learning models without worrying about the cost, skills, and infrastructure required to deploy AI applications globally. At Gcore, we believe that the edge is where you get the best performance and end-user experience, so we are continuously innovating to ensure all our customers get unparalleled scale and performance. Gcore Inference at the Edge gives you all the power without the hassle, delivering a modern, effective, and efficient AI inference experience.”

For more information, visit https://gcore.com/inference-at-the-edge.

About Gcore

Gcore is a global provider of Edge AI, Cloud, Network and Security solutions. Gcore provides solutions to global leaders across various industries. The company manages its own global IT infrastructure across six continents and boasts the best network performance in Europe, Africa and Latin America with an average response time of 30 milliseconds worldwide. Gcore's network consists of over 180 points of presence around the world in highly reliable Tier IV and Tier III data centers with a total capacity of over 200 Tbps.

contact address

Gcore Press Contact
[email protected]





Source link

Leave a Reply

Your email address will not be published. Required fields are marked *