Hacker Newsnew | past | comments | ask | show | jobs | submit | huy's commentslogin

Congrats on the launch!

I think here's a few players in this space (dev-friendly BI tool) already: - Holistics.io - Lightdash - Hashboard

These tools all allow analysts to use both/either a local/cloud IDE to write analytics logic, and check in to Git version control.

How do you plan to differentiate with them?


> While it looks really nice, hard to imagine it's quicker to hand draw this than reverse it out of your existing schema

They have an option to take your SQL dump text and convert it back to DBML which is the DSL they use to render.


I've tried doing the SQL dump on dbdiagram a few months ago and ran into some issues. Unfortunately I don't remember specifically what, but after testing out a few other programs I ended up using the free version of https://dbschema.com/ and have been very happy with it.


Came here to ask what the DSL was. Thanks for pointing that out.

Never heard of DBML before... looks somewhat interesting.

Any other DB DSLs worth recommending?


Ah perfect - did not see that! thanks


If you're new to the BI space, this book should give you a pretty good high-level overview: https://www.holistics.io/books/setup-analytics/


If you’re using Chartio and looking to migrate, here are a few alternatives you can consider. These are mostly cloud-based tool that is SQL-native and sends your written SQL queries to your SQL data-warehouse.

1/ Redash: write SQL gets charts, similar to Chartio (although a bit less polished as it was developed by a small team). Open-source or Cloud-based. Downside is they recently get acquired by Databricks, so I’m not sure how that will affect their product roadmap. Support for Google Sheets as well.

2/ Superset: also write SQL gets charts. They’re both open-source (Apache Incubated), and being supported by a commercial company behind (Preset). Their visualization library looks pretty diverse too.

3/ Metabase: This is also based on SQL. The difference is that they put a simple modeling layer in between your DW and SQL, so that they provide a simple ‘ask question’ interface for business users to get data, it’s quite simple though so if business users want more sophisticated questions they’ll need to resort back to SQL. Their UI is pretty sleek. Both open source and cloud-based (pricing starts from $100) with a commercial company behind.

4/ Holistics: Similar to Metabase, where they provide a modeling layer and self-service capability to non-technical layer. Difference is their modeling layer seems more sophisticated than that of Metabase. They also support the T in ELT as part of modeling. Support for Google Sheets as well. Their UI looks well-polished. Cloud-based only, with pricing starting from $100. They’re also working on Git version-controlling your SQL/modeling code so that it’s easier to track change/audit, this might help prevent vendor lock-in and ease future migration.

5/ Mode: Write SQL gets charts (similar to Redash/Superset). They’re a lot more polished than Redash, with a focus towards doing Data Science (native Python/R support, Notebook, etc). Their Free plan supports up to 5 users, but beyond that you will need to talk to their sales people to get quoted pricing. I heard it can get over $1k/month based with additional user pricing on top.

6/ Looker: probably best in-class in term of SQL-based data modeling, but very expensive.

7/ There are also other tools which I don’t think fit in the SQL-native narrative, like AWS Quicksight, Google Data Studio, PowerBI, Tableau and Qlik.

P/s: These 2 posts give a fairly good framework on how to look/evaluate different BI tools. This is a pretty crowded space with many different paradigms baked in over many years, so it might be a bit confusing for new people looking into the field

- https://www.holistics.io/books/setup-analytics/navigating-th...

- https://www.holistics.io/books/setup-analytics/the-arc-of-ad...


There are so many such solutions, that It depends on your use cases.

If you're looking for high-level dashboarding/monitoring that connects directly with your sources (REST, 3rd-party apps like Facebook, Adwords, etc), then tools in the realm of Klipfolio, Databox, Grow.com will suit you well. If you're looking for visualization tools that connect to your data-warehouse/SQL database, then BI tools like Holistics, Metabase, Redash, etc will likely work for you. And of course there's tools that can support both kinds of sources, like PowerBI and Google Data Studio.

P/s: This post does a good job at laying down a framework for evaluating these sort of BI tools: https://www.holistics.io/books/setup-analytics/navigating-th...


This is a great reply.

A nondev should likely stick to the first group, which has a lot of integrations and the hardest coding will be around excel formulas at most

In my experience its best to do the following:

Have 1 dashboard specific to business finance and ads metrics. This dashboard includes klipfolio or simpler things like grow or fathom accounting.

Another tool connected to your bd, at the most basic starting with mode Analytics and then going full visual with looker or tableau (this is where things get expensive)


You seem to be kind of shadowban (not sure what the proper term is), you might want to write an email to the hn moderation to clear that up


you were correct. Thanks for pointing this out.

Have a great new year!


Grow.com supports both 3rd party, as well as SQL database. It doesn't run SQL directly on your db, but pulls it into the provided data warehouse.

I work at Grow, happy to answer any questions.


For those who're curious, the tool to build the database diagrams is dbdiagram.io


I counted 6.

Fivetran, dbt, Preset (Superset/Airflow), Sisu, Imply and Databricks.

Though, as someone who's in this space a while, I think they did a decently fair job at articulating the 'modern' data infrastructure landscape.


I think you have a point, but there are more nuiances than that.

There are typically 2 types of data to collect: Transactional data and behavioural data.

Most transactional data, due to their important nature, are already generated and captured by the production applications. Since the logic is coded by application engineer, it's usually hard to get this data wrong. These data are then ETL-ed (or EL-ed) over to a DW, as described by the article.

For behavioural data, this is where your statement will most apply to. This is where tools like Snowplow, Posthog, Segment, etc come in to set up the proper event data collection engine. This is also where it's important to "collect data properly", as these kinds of event data changes structure fast, and hard to keep track over time. I'd admit this space (data collection management) is still nascent, with only tools like iterative.ly on the market.


For those who're interested in learning more about the history and evolution of data infrastructure/BI - basically why and how it has come to this stage - check out this short guidebook [1] that my colleagues and I put together a few months back.

It goes into details how much relevance the practices of the past (OLAP, Kimball's modeling) has with the current changes in by the cloud era (MPP, cheap storage/compute, etc). Chapter 4 will be most interesting for HN audience: It walks through the different waves of data adoption ever since BI was invented in the 60-70s.

https://holistics.io/books/setup-analytics/


This sounds like an in-depth discussion of what the a16z document calls Blueprint 1: Modern Business Intelligence. I don’t know if the other two blueprints for Multimodal and AI are explored.


I know of three (all OSS)

* https://www.posthog.com/ (more mixpanel alternative, can auto-track events too)

* https://snowplowanalytics.com/

* https://rudderstack.com/ (explicitly say it's Segment alternative)


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: