Alternatives to Kibana logo

Alternatives to Kibana

Datadog, Grafana, Loggly, Graylog, and Splunk are the most popular alternatives and competitors to Kibana.
20.1K
16K
+ 1
261

What is Kibana and what are its top alternatives?

Kibana is an open source (Apache Licensed), browser based analytics and search dashboard for Elasticsearch. Kibana is a snap to setup and start using. Kibana strives to be easy to get started with, while also being flexible and powerful, just like Elasticsearch.
Kibana is a tool in the Monitoring Tools category of a tech stack.
Kibana is an open source tool with 19.3K GitHub stars and 8K GitHub forks. Here’s a link to Kibana's open source repository on GitHub

Top Alternatives to Kibana

  • Datadog
    Datadog

    Datadog is the leading service for cloud-scale monitoring. It is used by IT, operations, and development teams who build and operate applications that run on dynamic or hybrid cloud infrastructure. Start monitoring in minutes with Datadog! ...

  • Grafana
    Grafana

    Grafana is a general purpose dashboard and graph composer. It's focused on providing rich ways to visualize time series metrics, mainly though graphs but supports other ways to visualize data through a pluggable panel architecture. It currently has rich support for for Graphite, InfluxDB and OpenTSDB. But supports other data sources via plugins. ...

  • Loggly
    Loggly

    It is a SaaS solution to manage your log data. There is nothing to install and updates are automatically applied to your Loggly subdomain. ...

  • Graylog
    Graylog

    Centralize and aggregate all your log files for 100% visibility. Use our powerful query language to search through terabytes of log data to discover and analyze important information. ...

  • Splunk
    Splunk

    It provides the leading platform for Operational Intelligence. Customers use it to search, monitor, analyze and visualize machine data. ...

  • Prometheus
    Prometheus

    Prometheus is a systems and service monitoring system. It collects metrics from configured targets at given intervals, evaluates rule expressions, displays the results, and can trigger alerts if some condition is observed to be true. ...

  • Tableau
    Tableau

    Tableau can help anyone see and understand their data. Connect to almost any database, drag and drop to create visualizations, and share with a click. ...

  • New Relic
    New Relic

    The world’s best software and DevOps teams rely on New Relic to move faster, make better decisions and create best-in-class digital experiences. If you run software, you need to run New Relic. More than 50% of the Fortune 100 do too. ...

Kibana alternatives & related posts

Datadog logo

Datadog

9.2K
7.9K
857
Unify logs, metrics, and traces from across your distributed infrastructure.
9.2K
7.9K
+ 1
857
PROS OF DATADOG
  • 137
    Monitoring for many apps (databases, web servers, etc)
  • 107
    Easy setup
  • 87
    Powerful ui
  • 83
    Powerful integrations
  • 70
    Great value
  • 54
    Great visualization
  • 46
    Events + metrics = clarity
  • 41
    Custom metrics
  • 41
    Notifications
  • 39
    Flexibility
  • 19
    Free & paid plans
  • 16
    Great customer support
  • 15
    Makes my life easier
  • 10
    Adapts automatically as i scale up
  • 9
    Easy setup and plugins
  • 8
    Super easy and powerful
  • 7
    AWS support
  • 7
    In-context collaboration
  • 6
    Rich in features
  • 5
    Docker support
  • 4
    Cost
  • 4
    Source control and bug tracking
  • 4
    Automation tools
  • 4
    Cute logo
  • 4
    Monitor almost everything
  • 4
    Full visibility of applications
  • 4
    Simple, powerful, great for infra
  • 4
    Easy to Analyze
  • 4
    Best than others
  • 3
    Expensive
  • 3
    Best in the field
  • 3
    Free setup
  • 3
    Good for Startups
  • 2
    APM
CONS OF DATADOG
  • 19
    Expensive
  • 4
    No errors exception tracking
  • 2
    External Network Goes Down You Wont Be Logging
  • 1
    Complicated

related Datadog posts

Noah Zoschke
Engineering Manager at Segment · | 30 upvotes · 267.1K views

We just launched the Segment Config API (try it out for yourself here) — a set of public REST APIs that enable you to manage your Segment configuration. Behind the scenes the Config API is built with Go , GRPC and Envoy.

At Segment, we build new services in Go by default. The language is simple so new team members quickly ramp up on a codebase. The tool chain is fast so developers get immediate feedback when they break code, tests or integrations with other systems. The runtime is fast so it performs great at scale.

For the newest round of APIs we adopted the GRPC service #framework.

The Protocol Buffer service definition language makes it easy to design type-safe and consistent APIs, thanks to ecosystem tools like the Google API Design Guide for API standards, uber/prototool for formatting and linting .protos and lyft/protoc-gen-validate for defining field validations, and grpc-gateway for defining REST mapping.

With a well designed .proto, its easy to generate a Go server interface and a TypeScript client, providing type-safe RPC between languages.

For the API gateway and RPC we adopted the Envoy service proxy.

The internet-facing segmentapis.com endpoint is an Envoy front proxy that rate-limits and authenticates every request. It then transcodes a #REST / #JSON request to an upstream GRPC request. The upstream GRPC servers are running an Envoy sidecar configured for Datadog stats.

The result is API #security , #reliability and consistent #observability through Envoy configuration, not code.

We experimented with Swagger service definitions, but the spec is sprawling and the generated clients and server stubs leave a lot to be desired. GRPC and .proto and the Go implementation feels better designed and implemented. Thanks to the GRPC tooling and ecosystem you can generate Swagger from .protos, but it’s effectively impossible to go the other way.

See more
Robert Zuber

Our primary source of monitoring and alerting is Datadog. We’ve got prebuilt dashboards for every scenario and integration with PagerDuty to manage routing any alerts. We’ve definitely scaled past the point where managing dashboards is easy, but we haven’t had time to invest in using features like Anomaly Detection. We’ve started using Honeycomb for some targeted debugging of complex production issues and we are liking what we’ve seen. We capture any unhandled exceptions with Rollbar and, if we realize one will keep happening, we quickly convert the metrics to point back to Datadog, to keep Rollbar as clean as possible.

We use Segment to consolidate all of our trackers, the most important of which goes to Amplitude to analyze user patterns. However, if we need a more consolidated view, we push all of our data to our own data warehouse running PostgreSQL; this is available for analytics and dashboard creation through Looker.

See more
Grafana logo

Grafana

17.4K
14K
415
Open source Graphite & InfluxDB Dashboard and Graph Editor
17.4K
14K
+ 1
415
PROS OF GRAFANA
  • 89
    Beautiful
  • 68
    Graphs are interactive
  • 57
    Free
  • 56
    Easy
  • 34
    Nicer than the Graphite web interface
  • 26
    Many integrations
  • 18
    Can build dashboards
  • 10
    Easy to specify time window
  • 10
    Can collaborate on dashboards
  • 9
    Dashboards contain number tiles
  • 5
    Open Source
  • 5
    Integration with InfluxDB
  • 5
    Click and drag to zoom in
  • 4
    Authentification and users management
  • 4
    Threshold limits in graphs
  • 3
    Alerts
  • 3
    It is open to cloud watch and many database
  • 3
    Simple and native support to Prometheus
  • 2
    Great community support
  • 2
    You can use this for development to check memcache
  • 2
    You can visualize real time data to put alerts
  • 0
    Grapsh as code
  • 0
    Plugin visualizationa
CONS OF GRAFANA
  • 1
    No interactive query builder

related Grafana posts

Conor Myhrvold
Tech Brand Mgr, Office of CTO at Uber · | 15 upvotes · 4.5M views

Why we spent several years building an open source, large-scale metrics alerting system, M3, built for Prometheus:

By late 2014, all services, infrastructure, and servers at Uber emitted metrics to a Graphite stack that stored them using the Whisper file format in a sharded Carbon cluster. We used Grafana for dashboarding and Nagios for alerting, issuing Graphite threshold checks via source-controlled scripts. While this worked for a while, expanding the Carbon cluster required a manual resharding process and, due to lack of replication, any single node’s disk failure caused permanent loss of its associated metrics. In short, this solution was not able to meet our needs as the company continued to grow.

To ensure the scalability of Uber’s metrics backend, we decided to build out a system that provided fault tolerant metrics ingestion, storage, and querying as a managed platform...

https://eng.uber.com/m3/

(GitHub : https://github.com/m3db/m3)

See more
Matt Menzenski
Senior Software Engineering Manager at PayIt · | 15 upvotes · 993K views

Grafana and Prometheus together, running on Kubernetes , is a powerful combination. These tools are cloud-native and offer a large community and easy integrations. At PayIt we're using exporting Java application metrics using a Dropwizard metrics exporter, and our Node.js services now use the prom-client npm library to serve metrics.

See more
Loggly logo

Loggly

274
302
168
Unified log analysis & log monitoring
274
302
+ 1
168
PROS OF LOGGLY
  • 37
    Centralized log management
  • 25
    Easy to setup
  • 21
    Great filtering
  • 16
    Live logging
  • 15
    Json log support
  • 10
    Log Management
  • 10
    Alerting
  • 7
    Great Dashboards
  • 7
    Love the product
  • 4
    Heroku Add-on
  • 2
    Easy to setup and use
  • 2
    Easy setup
  • 2
    No alerts in free plan
  • 2
    Great UI
  • 2
    Good parsing
  • 2
    Powerful
  • 2
    Fast search
  • 2
    Backup to S3
CONS OF LOGGLY
  • 3
    Pricey after free plan

related Loggly posts

Graylog logo

Graylog

575
704
70
Open source log management that actually works
575
704
+ 1
70
PROS OF GRAYLOG
  • 19
    Open source
  • 13
    Powerfull
  • 8
    Well documented
  • 6
    Alerts
  • 5
    User authentification
  • 5
    Flexibel query and parsing language
  • 3
    User management
  • 3
    Easy query language and english parsing
  • 3
    Alerts and dashboards
  • 2
    Easy to install
  • 1
    A large community
  • 1
    Manage users and permissions
  • 1
    Free Version
CONS OF GRAYLOG
  • 1
    Does not handle frozen indices at all

related Graylog posts

Splunk logo

Splunk

597
998
20
Search, monitor, analyze and visualize machine data
597
998
+ 1
20
PROS OF SPLUNK
  • 3
    API for searching logs, running reports
  • 3
    Alert system based on custom query results
  • 2
    Dashboarding on any log contents
  • 2
    Custom log parsing as well as automatic parsing
  • 2
    Ability to style search results into reports
  • 2
    Query engine supports joining, aggregation, stats, etc
  • 2
    Splunk language supports string, date manip, math, etc
  • 2
    Rich GUI for searching live logs
  • 1
    Query any log as key-value pairs
  • 1
    Granular scheduling and time window support
CONS OF SPLUNK
  • 1
    Splunk query language rich so lots to learn

related Splunk posts

Shared insights
on
SplunkSplunkDjangoDjango

I am designing a Django application for my organization which will be used as an internal tool. The infra team said that I will not be having SSH access to the production server and I will have to log all my backend application messages to Splunk. I have no knowledge of Splunk so the following are the approaches I am considering: Approach 1: Create an hourly cron job that uploads the server log file to some Splunk storage for later analysis. - Is this possible? Approach 2: Is it possible just to stream the logs to some splunk endpoint? (If yes, I feel network usage and communication overhead will be a pain-point for my application)

Is there any better or standard approach? Thanks in advance.

See more
Shared insights
on
KibanaKibanaSplunkSplunkGrafanaGrafana

I use Kibana because it ships with the ELK stack. I don't find it as powerful as Splunk however it is light years above grepping through log files. We previously used Grafana but found it to be annoying to maintain a separate tool outside of the ELK stack. We were able to get everything we needed from Kibana.

See more
Prometheus logo

Prometheus

4.1K
3.8K
239
An open-source service monitoring system and time series database, developed by SoundCloud
4.1K
3.8K
+ 1
239
PROS OF PROMETHEUS
  • 47
    Powerful easy to use monitoring
  • 38
    Flexible query language
  • 32
    Dimensional data model
  • 27
    Alerts
  • 23
    Active and responsive community
  • 22
    Extensive integrations
  • 19
    Easy to setup
  • 12
    Beautiful Model and Query language
  • 7
    Easy to extend
  • 6
    Nice
  • 3
    Written in Go
  • 2
    Good for experimentation
  • 1
    Easy for monitoring
CONS OF PROMETHEUS
  • 12
    Just for metrics
  • 6
    Bad UI
  • 6
    Needs monitoring to access metrics endpoints
  • 4
    Not easy to configure and use
  • 3
    Supports only active agents
  • 2
    Written in Go
  • 2
    TLS is quite difficult to understand
  • 2
    Requires multiple applications and tools
  • 1
    Single point of failure

related Prometheus posts

Conor Myhrvold
Tech Brand Mgr, Office of CTO at Uber · | 15 upvotes · 4.5M views

Why we spent several years building an open source, large-scale metrics alerting system, M3, built for Prometheus:

By late 2014, all services, infrastructure, and servers at Uber emitted metrics to a Graphite stack that stored them using the Whisper file format in a sharded Carbon cluster. We used Grafana for dashboarding and Nagios for alerting, issuing Graphite threshold checks via source-controlled scripts. While this worked for a while, expanding the Carbon cluster required a manual resharding process and, due to lack of replication, any single node’s disk failure caused permanent loss of its associated metrics. In short, this solution was not able to meet our needs as the company continued to grow.

To ensure the scalability of Uber’s metrics backend, we decided to build out a system that provided fault tolerant metrics ingestion, storage, and querying as a managed platform...

https://eng.uber.com/m3/

(GitHub : https://github.com/m3db/m3)

See more
Matt Menzenski
Senior Software Engineering Manager at PayIt · | 15 upvotes · 993K views

Grafana and Prometheus together, running on Kubernetes , is a powerful combination. These tools are cloud-native and offer a large community and easy integrations. At PayIt we're using exporting Java application metrics using a Dropwizard metrics exporter, and our Node.js services now use the prom-client npm library to serve metrics.

See more
Tableau logo

Tableau

1.2K
1.3K
8
Tableau helps people see and understand data.
1.2K
1.3K
+ 1
8
PROS OF TABLEAU
  • 6
    Capable of visualising billions of rows
  • 1
    Intuitive and easy to learn
  • 1
    Responsive
CONS OF TABLEAU
  • 2
    Very expensive for small companies

related Tableau posts

Looking for the best analytics software for a medium-large-sized firm. We currently use a Microsoft SQL Server database that is analyzed in Tableau desktop/published to Tableau online for users to access dashboards. Is it worth the cost savings/time to switch over to using SSRS or Power BI? Does anyone have experience migrating from Tableau to SSRS /or Power BI? Our other option is to consider using Tableau on-premises instead of online. Using custom SQL with over 3 million rows really decreases performances and results in processing times that greatly exceed our typical experience. Thanks.

See more
Shared insights
on
TableauTableauQlikQlikPowerBIPowerBI

Hello everyone,

My team and I are currently in the process of selecting a Business Intelligence (BI) tool for our actively developing company, which has over 500 employees. We are considering open-source options.

We are keen to connect with a Head of Analytics or BI Analytics professional who has extensive experience working with any of these systems and is willing to share their insights. Ideally, we would like to speak with someone from companies that have transitioned from proprietary BI tools (such as PowerBI, Qlik, or Tableau) to open-source BI tools, or vice versa.

If you have any contacts or recommendations for individuals we could reach out to regarding this matter, we would greatly appreciate it. Additionally, if you are personally willing to share your experiences, please feel free to reach out to me directly. Thank you!

See more
New Relic logo

New Relic

20.7K
8.5K
1.9K
New Relic is the industry’s largest and most comprehensive cloud-based observability platform.
20.7K
8.5K
+ 1
1.9K
PROS OF NEW RELIC
  • 415
    Easy setup
  • 344
    Really powerful
  • 244
    Awesome visualization
  • 194
    Ease of use
  • 151
    Great ui
  • 107
    Free tier
  • 80
    Great tool for insights
  • 66
    Heroku Integration
  • 55
    Market leader
  • 49
    Peace of mind
  • 21
    Push notifications
  • 20
    Email notifications
  • 17
    Heroku Add-on
  • 16
    Error Detection and Alerting
  • 13
    Multiple language support
  • 11
    Server Resources Monitoring
  • 11
    SQL Analysis
  • 9
    Transaction Tracing
  • 8
    Azure Add-on
  • 8
    Apdex Scores
  • 7
    Detailed reports
  • 7
    Analysis of CPU, Disk, Memory, and Network
  • 6
    Application Response Times
  • 6
    Performance of External Services
  • 6
    Application Availability Monitoring and Alerting
  • 6
    Error Analysis
  • 5
    JVM Performance Analyzer (Java)
  • 5
    Most Time Consuming Transactions
  • 4
    Top Database Operations
  • 4
    Easy to use
  • 4
    Browser Transaction Tracing
  • 3
    Application Map
  • 3
    Weekly Performance Email
  • 3
    Custom Dashboards
  • 3
    Pagoda Box integration
  • 2
    App Speed Index
  • 2
    Easy to setup
  • 2
    Background Jobs Transaction Analysis
  • 1
    Time Comparisons
  • 1
    Access to Performance Data API
  • 1
    Super Expensive
  • 1
    Team Collaboration Tools
  • 1
    Metric Data Retention
  • 1
    Metric Data Resolution
  • 1
    Worst Transactions by User Dissatisfaction
  • 1
    Real User Monitoring Overview
  • 1
    Real User Monitoring Analysis and Breakdown
  • 1
    Free
  • 1
    Best of the best, what more can you ask for
  • 1
    Best monitoring on the market
  • 1
    Rails integration
  • 1
    Incident Detection and Alerting
  • 0
    Cost
  • 0
    Exceptions
  • 0
    Price
  • 0
    Proce
CONS OF NEW RELIC
  • 20
    Pricing model doesn't suit microservices
  • 10
    UI isn't great
  • 7
    Expensive
  • 7
    Visualizations aren't very helpful
  • 5
    Hard to understand why things in your app are breaking

related New Relic posts

Cooper Marcus
Director of Ecosystem at Kong Inc. · | 17 upvotes · 110.1K views
Shared insights
on
New RelicNew RelicGitHubGitHubZapierZapier
at

I've used more and more of New Relic Insights here in my work at Kong. New Relic Insights is a "time series event database as a service" with a super-easy API for inserting custom events, and a flexible query language for building visualization widgets and dashboards.

I'm a big fan of New Relic Insights when I have data I know I need to analyze, but perhaps I'm not exactly sure how I want to analyze it in the future. For example, at Kong we recently wanted to get some understanding of our open source community's activity on our GitHub repos. I was able to quickly configure GitHub to send webhooks to Zapier , which in turn posted the JSON to New Relic Insights.

Insights is schema-less and configuration-less - just start posting JSON key value pairs, then start querying your data.

Within minutes, data was flowing from GitHub to Insights, and I was building widgets on my Insights dashboard to help my colleagues visualize the activity of our open source community.

#GitHubAnalytics #OpenSourceCommunityAnalytics #CommunityAnalytics #RepoAnalytics

See more
Julien DeFrance
Principal Software Engineer at Tophatter · | 16 upvotes · 3.1M views

Back in 2014, I was given an opportunity to re-architect SmartZip Analytics platform, and flagship product: SmartTargeting. This is a SaaS software helping real estate professionals keeping up with their prospects and leads in a given neighborhood/territory, finding out (thanks to predictive analytics) who's the most likely to list/sell their home, and running cross-channel marketing automation against them: direct mail, online ads, email... The company also does provide Data APIs to Enterprise customers.

I had inherited years and years of technical debt and I knew things had to change radically. The first enabler to this was to make use of the cloud and go with AWS, so we would stop re-inventing the wheel, and build around managed/scalable services.

For the SaaS product, we kept on working with Rails as this was what my team had the most knowledge in. We've however broken up the monolith and decoupled the front-end application from the backend thanks to the use of Rails API so we'd get independently scalable micro-services from now on.

Our various applications could now be deployed using AWS Elastic Beanstalk so we wouldn't waste any more efforts writing time-consuming Capistrano deployment scripts for instance. Combined with Docker so our application would run within its own container, independently from the underlying host configuration.

Storage-wise, we went with Amazon S3 and ditched any pre-existing local or network storage people used to deal with in our legacy systems. On the database side: Amazon RDS / MySQL initially. Ultimately migrated to Amazon RDS for Aurora / MySQL when it got released. Once again, here you need a managed service your cloud provider handles for you.

Future improvements / technology decisions included:

Caching: Amazon ElastiCache / Memcached CDN: Amazon CloudFront Systems Integration: Segment / Zapier Data-warehousing: Amazon Redshift BI: Amazon Quicksight / Superset Search: Elasticsearch / Amazon Elasticsearch Service / Algolia Monitoring: New Relic

As our usage grows, patterns changed, and/or our business needs evolved, my role as Engineering Manager then Director of Engineering was also to ensure my team kept on learning and innovating, while delivering on business value.

One of these innovations was to get ourselves into Serverless : Adopting AWS Lambda was a big step forward. At the time, only available for Node.js (Not Ruby ) but a great way to handle cost efficiency, unpredictable traffic, sudden bursts of traffic... Ultimately you want the whole chain of services involved in a call to be serverless, and that's when we've started leveraging Amazon DynamoDB on these projects so they'd be fully scalable.

See more