Security and Compliance
A brief summary of Hyperswitch infrastructure ensuring compliance and data privacy
Data center locations
Hyperswitch Cloud is hosted on the AWS platform and served out of the US (N.Virginia region) by default. Hyperswitch is listed partner on AWS Partner Network.
We support deployments in more geographies for merchants who wish to restrict data transfer across geographies. Each deployment is supported with at least two redundant Availability Zones for ensuring reliability.
Global Server (US) | EU Residency Server | |
---|---|---|
Live endpoints |
| Available on request |
Note: Hyperswitch is a super lightweight payment switch with Infrastructure-as-Code capability. So we can quickly deploy in new geographies (AWS regions) as data residency requirements emerge across countries around the world.
Latency
The application latency of Hyperswitch is optimized to sub-30 milliseconds
(90 percentile) to ensure that the application by itself does not add any latency cost. Read here to know more about it.
Inter-regional network latency is optimized by fronting the API endpoints with Cloud Delivery Network (AWS Cloudfront).
Note: In case you are opting for the Hyperswitch Open Source offering, you will be able to run the software like a microservice within your own deployment.
Reliability
Reliability is ensured as a primitive in each component of Hyperswitch to offer 99.99% uptime guarantee.
All infrastructure components are deployed across multiple Availability Zones. This includes:
Load Balancers
Proxy instances for incoming and outgoing APIs
Kubernetes clusters powering core application, analytics and async services
Aurora PSQL Database (primary storage)
Elasticache storage (cache)
Vault service
Scalability
Hyperswitch support for 80 RPS per merchant account and rate limits applicable beyond. Higher enterprise workloads can be processed on request basis.
Scaling Infrastructure components
Scaling of each component in the Hyperswitch setup is achieved as follows
Incoming proxy layer is deployed using Auto Scaling Group (ASG)
Application layer is managed using Kubernetes with Horizontal Pod Autoscaler (HPA)
Outgoing Proxy is deployed using Auto Scaling Group (ASG)
Handling traffic spikes
Database provides consistency, but it more often becomes a bottleneck and a failure point in case of sudden traffic spikes. Hyperswitch can be configured (on request basis) to use Cache as Data storage in case of sudden spikes in the traffic to reduce the friction of database latency. All the data is drained from cache to the database.
Last updated