This highly depends on the project. We do have pipelines on Azure Data Factory. Code is implemented as a library as well as some Databricks notebooks are running on production. We also have some python processes being run on Kubernetes.
Our biggest data set is IoT data of hundreds of thousands devices our end-users have connected to the cloud. It results in petabytes of data that we analyse, e.g. to provide remote service algorithms. We also support Supply Chain Management topics (i.e. production, logistics, sales) where we provide algorithms supporting planning and provide analysis services to detect and resolve any bottlenecks.
Rows are being added to the largest data set in this domain every day
Roles in the projects:
A CARRIER COMPANY - © 2024 CARRIER