Frequently Asked Questions
Apica is a data platform for engineers built using cloud-native technologies. Using Apica, you can unify data from your engineering teams and aggregate logs, metrics, and traces, connect APIs, databases, and data lakes, and analyse them, all through a single interface. Apica uses S3 as the primary data layer and makes data stored within S3 indexable and searchable. This gives you the ability to log as much as your IT ecosystem demands and store data for as long as you wish to, without the massive storage tax that other platforms levy for log data retention.
Apica isn’t just a log management platform but a complete observability solution that helps you dive deep into the unknowns of your entire technology stack, all in one platform.
With our log management capabilities, you can stream and analyze logs across your applications and infrastructure in real-time, generate quick and insightful visualizations from log data, and troubleshoot issues with AI and ML-driven analytics at any scale.
Apica’s Prometheus-backed, monitoring engine generates insightful metrics that lets you monitor the health and performance of all your applications and infrastructure and troubleshoot anomalies at scale.
Apica’s built-in SIEM rules engine detects anomalous security events in real-time and at an infinite scale. You can sample security events in log data against crowdsourced security rules from the Sigma project for faster threat detection.
Moreover, you can analyze all your APIs, identify patterns and behavior that help you troubleshoot issues, prevent threats preemptively, understand API usage, and confidently ship great APIs.
Apica SaaS can be set up and be ready for use within a few minutes. Apica PaaS deployments may take longer depending on where you’d like to deploy it and the data sources you’d like to connect to Apica.
You can retain ingested data for as long as you wish. Since Apica uses S3 as the primary data store, you get the flexibility to log and store as much log data as you’d like to, without any storage tax.
Both Apica SaaS and Apica PaaS have no limit on the number of users. You can onboard as many users as you’d like and manage them using fine-grained RBAC policies directly from the Apica UI. Several of our existing customers have more than 100 active users of Apica each.
For an ingest volume of 100 GB per day, the minimum hardware requirements are 16-core vCPU with 32 GB RAM. If you’re looking at ingesting higher volumes, you can get in touch with our Support Team for accurate sizing.
We do not follow the practice of surge pricing. We provision for peak data rates exceeding which the log forwarder slows down and Apica starts rejecting ingestion requests. The log forwarder resumes forwarding once the data rates fall well below the threshold.
Our pricing is simple and straightforward. Your daily log ingestion volume is the only factor that affects pricing. For more information on plans and pricing, refer to the Pricing page.
Absolutely! You can run multiple clusters of Apica that you can easily manage via the Apica UI.
Yes! Apica PaaS comes with a free-forever Community Edition that includes 4 ingest workers. You can easily deploy the Community Edition on a Kubernetes cluster using the Apica PaaS Helm Chart. You can find detailed instructions on deploying Apica PaaS here: https://docs.apica.io/Apica-server/Apica-paas-community-edition
At the end of the 14-day free trial period, you’ll be asked whether you’d like to continue with the plan you selected while signing up, or upgrade to another plan that suits your needs better. The payment method you selected during sign-up will be billed as per the plan you select. In case you wish to cancel your subscription, reach out to us and we’ll take care of it.
To promote maximum flexibility, Apica can be hosted as a self-service PaaS or managed SaaS solution.
Users of solutions that rely on Logstash or Fluentd for ingesting logs, such as the ELK stack, Logz, Logentries, etc., can switch over to Apica by simply reconfiguring Logstash or Fluentd to forward logs to Apica instead.
Other solutions like Splunk, Sumo Logic, or Datadog use their own proprietary log collectors. If you use any of these solutions, we can switch you over to Apica by setting up Logstash, Fluentd, or our very own LogFlow to handle log forwarding duties.
Yes. Containerization was a major consideration when developing Apica. Apica is a solution focusing on modern deployment options, with integrations and existing configurations designed for Kubernetes.
Apica currently supports integrations with popular log collection and forwarding agents such as Logstash, Rsyslog, Fluentd, and Fluent Bit. You can also use the Docker syslog driver to connect your container applications to Apica. Besides, you can also use Logflow to connect to any data endpoint, whether cloud, edge, or on-premise, to collect, transform, and ship logs to Apica.
Absolutely! You can forward your syslog data to a syslog server such as Rsyslog. You can then forward this data from Rsyslog using omfwd or omrelp to your Apica instance via an RELP-aware endpoint. Alternatively, you can leverage Logstash’s Syslog out plugin to funnel and forward syslog data into Apica.
Yes. Along with native integrations with alert platforms like Email, Slack, PagerDuty, OpsGenie, Mattermost, HipChat, and ChatWork, Apica also supports integrating with custom webhooks for forwarding alerts triggered within Apica. You can also work with our Support Team for bespoke integrations with your existing notification system.