I am a Databricks service partner, and my customers use Azure Databricks and Data Factory.
Databricks Data Intelligence Platform
Databricks, Inc.External reviews
External reviews are not included in the AWS star rating for the product.
Lake House platform review
A Robust Solution for Big Data Management and Analytics
Experience with databricks
End-to-end support for machine learning and faster AI delivery.
Databricks - A breath of Fresh air in Big Data
The platform has a solution for every data person, including but not limited to a Notebook that works with Scala, Python, R and SQL, a traditional SQL Editor, downloadable datasets and in house visualisations just a click away!
Data Lake but combined with Datawarehouse benefits
This solution has eliminated dependency on our already saturated datawarehouse resources. This has also helped in debugging as all data is processed and resides in one place.Last but not the least, this has reduced costs of our datawarehouse by 20%
Databricks Genie Code - Agentic Applied AI for end-end SDL liefecycle
1) Genie Code automated our ETL processes, reducing manual effort and increasing efficiency. With Agentic’s SDL, we implemented CI/CD pipelines for faster, seamless updates and deployments.
2) Genie Code streamlined complex STTM mappings, improving accuracy and speed. Agentic’s real-time updates ensured mapping adjustments were made dynamically to align with changing transaction data.
3) We defined automated unit tests using SKILL.md, ensuring data transformations are validated before deployment. This reduced errors and ensured data quality, boosting confidence in our analytics.
4) Using Skills.md, we added custom extensions to Genie Code, such as integrating third-party data for enriched reports. This agility allowed us to quickly adapt to business needs and deliver new capabilities.
5) Agentic’s SDL enabled real-time data processing, providing immediate analytics for decision-making. Our marketing and sales teams now act on fresh data instantly, improving response times and overall efficiency.
Debugging issues in complex workflows can be time-consuming due to limited visibility into intermediate data transformations.
Genie Code lacks advanced error recovery mechanisms, making it difficult to manage failures in large-scale data pipelines.
As data volume increases, Genie Code’s performance can degrade, requiring significant manual adjustments to ensure smooth operation at scale.
2) Genie Code automates end-to-end ETL workflows, from data extraction to transformation and loading, streamlining data operations and eliminating manual tasks.
3) Real time collaboration - Genie Code enables real-time collaboration across teams by using shared notebooks, making it easier for data professionals to build and refine workflows collectively.
A Tool Box to the Modern Big Data Data Scientist
Great tool for data exploration and development, no so much for production pipelines
Shareability
Hard to incorporate without being databricks aware, which leads to a vendor lock
Developing spark jobs towards production
A powerful solution that is easily integrated into a variety of platforms
What is our primary use case?
What is most valuable?
It's very simple to use Databricks Apache Spark. It's really good for parallel execution to scale up the workload. In this context, the usage is more about virtual machines.
Using meta-stores like Hive was optional, and the solution is good for data science use cases. With the Authenticator Log, Databricks is good for data transformation and BI usage. We have a platform.
What needs improvement?
I would like more integration with SQL for using data in different workspaces. We use the user interface for some functionalities, while for others, we have to use SQL to create data sets and grant permissions. For example, when creating a cluster, we have to create it with some API or user interface. Creating a cluster with some properties using SQL grants the possibility of using SQL syntax. Integration with SQL will make Databricks easier to use by people who have experience with databases like Lakehouse, and they would be able to use the data lake and BI. More integration will help have one point of view for everyone using SQL syntax.
Integration with Kubernetes could also be good for minimizing the price because you can use Kubernetes instead of virtual machines. But that won't be easy.
For how long have I used the solution?
I have worked with the solution for four or five years, with some experience since 2016.
What do I think about the stability of the solution?
The solution is stable. The only problem with stability would be that people are not using it efficiently.
What do I think about the scalability of the solution?
The solution is good for scalability.
How was the initial setup?
When we have administration experience, the solution is not difficult to deploy. Technically, however, it's difficult because governance is more complex. For example, I have two warehouses on Databricks, which are clusters in this workspace, and we have to switch from workspace to workspace to have all this information. There is a system table that has all this, but I don't know if everyone can use these tables.
What's my experience with pricing, setup cost, and licensing?
Databricks are not costly when compared with other solutions' prices.
Which other solutions did I evaluate?
What other advice do I have?
People sometimes do not use the solution efficiently. They misunderstand databases, the usage of tables, and the performance. Many data engineers are very junior and don't have skills in that. Stability is more a customer problem than a problem with the product itself. One possible problem with the product is that there's no method to pause the usage of something. For example, we have to use the meta server or the data catalog in Synapse. But in Databricks, we have a choice to use a catalog or not, or Hive, which is always integrated, but we have to choose whether to use it or not. Many customers directly use the passes on Databricks, which causes performance and governance problems.
I can offer a lot of advice on Databricks, and one is to use meta stores like Unity Catalog or Hive Metastore. For incoming use cases, it's better to use Unity Catalog.
I rate Databricks a nine out of ten.
Processes large data for data science and data analytics purposes
What is our primary use case?
It's mainly used for data science, data analytics, visualization, and industrial analytics.
What is most valuable?
Specifically for data science and data analytics purposes, it can handle large amounts of data in less time. I can compare it with Teradata. If a job takes five hours with Teradata databases, Databricks can complete it in around three to three and a half hours.
So that's why it's quite convenient to use for data science, for training machine learning models. By using more computing power, you can make it even faster.
What needs improvement?
There is room for improvement in visualization.
For how long have I used the solution?
I used it for two years. I worked with the latest update.
What do I think about the stability of the solution?
I would rate the stability a nine out of ten. I didn't face performance drops.
What do I think about the scalability of the solution?
I would rate the scalability an eight out of ten.
How are customer service and support?
Databrick's support is great. If we need any support, they are very quick with it. And they genuinely want you to use Databricks. So, whatever we ask them, they come up with multiple solutions to problem statements. That's really good.
Overall, the customer service and support are very good.
Which solution did I use previously and why did I switch?
I personally prefer using Databricks. However, we also considered using Snowflake, but the pricing was different. It's price per query.
So, as per your storage, a data scientist or a data analytics team needs to query again and again, which does not suit a data-heavy organization.
What was our ROI?
It's a good return on investment for Databricks from a delivery perspective. Delivered multiple dashboards. So, it's quite a good return on investment. And being a small organization, everyone can use Databricks, and cost-wise, it's also good for small organizations.
Which other solutions did I evaluate?
If the company is a startup, Databricks might be suitable. If a big company needs a lot of storage, Teradata might be best for them. It depends on the situation.
What other advice do I have?
Overall, I would rate the solution a eight out of ten. I would definitely recommend this solution for small organizations.