Datalake is the central place to store big data from multiple sources.

Datalake can be used to store structured, semi-structured, or unstructured data.

Datalake is usually deployed on commodity hardware and can scale easily as per the user's need.

Datalake can be the setup of cloud and on-premises as per the requirement.

Managing datalake on on-premises requires the team to integrate lots of complex components but it gives users the flexibility to customise as per their needs.

Cloud data lake is a great solution if you do need want to set up your infrastructure. However, it is very costly as you need to pay for storage, bandwidth, etc.

The key components of datalake are Hadoop, spark,Hbase, MongoDB, Minio, druid,etc