
Gcore, a leading provider of edge AI, cloud, networking, and security solutions, has unveiled its latest offering: Inference at the Edge. This innovative solution promises to revolutionize the deployment of AI applications by enabling ultra-low latency processing at the edge, where data is generated and action is needed in real-time.
With Gcore Inference at the Edge, companies spanning various sectors—from manufacturing to automotive, retail, and technology—can deploy AI models swiftly, cost-effectively, and securely. This opens doors to a plethora of use cases including generative AI, object recognition, behavioral analytics, virtual assistants, and production monitoring, all accessible globally and with minimal latency.
Operating on Gcore’s extensive network of over 180 edge nodes interconnected via advanced low-latency routing technology, Inference at the Edge leverages NVIDIA’s cutting-edge L40S GPUs dedicated to AI inference tasks. These strategically positioned nodes ensure that data processing is as close to end-users as possible, with response times typically under 30 milliseconds.
The platform supports a diverse array of core ML models as well as custom ones, available through the Gcore ML Model Hub. This flexibility allows for on-demand selection and training of models before distribution across the global network, addressing the common issue of performance degradation when running models on the same servers used for training.
Beyond its technical capabilities, Gcore Inference at the Edge offers several advantages:
- Cost-effective Deployment: A flexible pricing structure ensures clients only pay for utilized resources.
- Built-in DDoS Protection: ML endpoints are automatically shielded from DDoS attacks, enhancing security.
- Top-tier Data Protection: GDPR compliance, alongside adherence to PCI DSS and ISO/IEC 27001 standards, ensures robust data security.
- Model Autoscaling: Models can auto-scale to handle peak loads, ensuring seamless operation even under sudden spikes in demand.
- Unlimited Object Storage: Scalable cloud storage enables flexible, reliable, and cost-effective solutions adaptable to evolving model needs.
Andre Reitenbach, Managing Director of Gcore, emphasizes the customer-centric approach: “With Gcore Inference at the Edge, our clients can focus on refining their ML models, free from concerns about costs, capacity, and infrastructure needed for global AI deployment. We believe that optimal performance and user experience lie at the network edge, and we’re committed to delivering unrivaled scalability and performance to every user.”
Gcore Inference at the Edge is now available, offering a powerful yet user-friendly solution for modern real-time data processing through AI. Learn more at: https://gcore.com/inference-at-the-edge