Categories
OAC direct connector with Cloud Storage Buckets

Organization Name
Genpact
Description
Need direct connection from OAC with Cloud Storage Buckets (OCI object storage, S3, Azure Data lake Storage etc.). This way dataset’s can be directly consumed.
Use Case and Business Need
With modern data lake solutions most of the data get stored in the storage bucket during ingestion and post processing. If OAC can have direct connector with cloud storage buckets like OCI object storage, S3, Azure Data lake Storage etc. Data will be easily accessible for analytics needs.
Original Idea Number: 736511779d
Comments
-
Wish this is enabled. Save processing times.
1 -
Since OAC has the ability to profile data using Explain, it would be very helpful to be able to analyze data that has been uploaded into Oracle Object Storage before deciding how much of it should be loaded into ADW. This is especially helpful when the data is large so that time and effort of loading it into ADW can be optimized.
0 -
This would save customers from having to move and transform data in order to report on it. Many customers have data in many locations beyond a database. Other competitor BI tools allow direct reads against object store data sets.
0 -
Since there's a option to connect to Google Drive or Dropbox, it makes sense to have native connectors to OCI Object Storage [AWS S3, Azure Blob].
0 -
Would be nice to use the object storage as a direct source.
0 -
As Object storage is natural store for image, logs - this must be a standard feature for any analytics, big data use case.
It is high time that this feature is made available in OAC. My one of Swedish retailer is stuck with issue now!
Cheers
Ajit
0 -
+1 to this one.
I have a business customer who uses Azure Blob storage as pretty much their data warehouse and are very happy with Power BI's ability to connect and work with files as if they are database tables. This ability would greatly accelerate OAC's adoption. Not to mention, the same API could be used to store snapshots on object storage as opposed to downloading them.
0 -
The detail that is generally missing from this discussion is a description of what kind of object/data you expect to use that is stored in the object storage. Are you talking about xlsx and csv that Data Sets can consume or something else? If it's something else, how do you intend to use it?
0 -
Hi Adam,
This is my customer's use case:
Data comes from various sources into Azure blob storage. They use Databricks Delta Lake (https://databricks.com/product/delta-lake-on-databricks) - pyspark to process data using spark jobs (dataframes within spark). At the end of the process, usable data is available as parquet files for PowerBI.
Simply put:
Data Sources --> Azure Blob Storage via Databricks Delta Lake --> Power BI
The need of the hour is to build a semantic layer on top of this data in blob storage. Just like on a database, they are looking for OAC to have the ability to ultimately make subject areas available on this data.
0 -
So what you are asking for is support for Parquet files (in addition to the existing xlsx and csv support), plus a connector to Azure Blob Storage and Oracle Object Storage?
0