Datalake is the central place to
big data from multiple sources.
Datalake can be used to
structured, semi-structured, or unstructured data.
Datalake is usually deployed on
hardware and can scale easily as per the user's need.
Datalake can be the
of cloud and on-premises as per the requirement.
Managing datalake on
requires the team to integrate lots of complex components but it gives users the flexibility to customise as per their needs.
lake is a great solution if you do need want to set up your infrastructure. However, it is very costly as you need to pay for storage, bandwidth, etc.
of datalake are Hadoop, spark,Hbase, MongoDB, Minio, druid,etc
7 things you should know about apache kafka