azure-data-factory

Azure Data Factory : querying _ts (epoch) field in CosmosDB source

孤街浪徒 提交于 2019-12-24 01:07:29
问题 I am trying to copy data from Azure CosmosDB into Azure SQL Database. In the Azure CosmosDB source, the query is as follows : select * from c where c.DefinitionTime >= '@{formatDateTime(pipeline().parameters.windowStart, 'yyyy-MM-ddTHH:mm:ssZ' )}' AND c.DefinitionTime < '@{formatDateTime(pipeline().parameters.windowEnd, 'yyyy-MM-ddTHH:mm:ssZ' )}' In CosmosDB, "DefinitionTime" is stored as a string. To the run the above query, I have to update "DefinitionTime" with range-indexing which is one

Authorization Time Out : Azure Data Lake Store Configuration in Azure Data Factory

≯℡__Kan透↙ 提交于 2019-12-24 00:34:49
问题 We are moving the data from Azure Table Storage to Azure Data Lake Store with the help of Data Factory. As part of if I have added a linked service which connects to Azure Data Lake Store. { "name": "AzureDataLakeStoreLinkedService", "properties": { "description": "", "hubName": "xxxxxxxxxxxxxxxxxxxxxx", "type": "AzureDataLakeStore", "typeProperties": { "dataLakeStoreUri": "https://xxxxxxxxxxxxx.azuredatalakestore.net/webhdfs/v1", "authorization": "**********", "sessionId": "**********",

What's reformatting my input data before I get to it?

时光怂恿深爱的人放手 提交于 2019-12-23 18:06:28
问题 I have a Data Lake Store account. I have a directory full of files containing data in JSON format, including some string values that contain times in ISO 8601 format, to wit: { "reading_time": "2008-09-15T15:53:00.91077" } Now when I create a Pipeline with a Data Factory that uses these JSON files as an input dataset, it sees the value of reading_time in a typical US format: "9/15/2008 3:53:00 PM" . Specifically, I get this message when I try to populate a DateTime field in the output dataset

How to perform Lookups in Azure Data Factory?

会有一股神秘感。 提交于 2019-12-23 15:46:23
问题 I'm a SSIS Developer. I do lots of SQL stored procedure lookup concepts in SSIS. But when coming to Azure Data Factory I haven't any idea how to perform a lookup using a SQL stored procedure. Could anyone please guide me on this? Thanks in advance ! Jay 回答1: Azure Data Factory (ADF) is more of an ELT tool rather than ETL, therefore direct lookups are not supported. Instead, this type of operation, along with other transforms is pushed down into the compute you are actually using. For example,

Disable activity in Azure Data factory pipeline without removing it

谁说胖子不能爱 提交于 2019-12-23 15:28:03
问题 So I am testing each of the activities of the pipeline and I want to disable some of the activities in it. Essentially there is an activity of sending emails which I want to disable as I wanted to see the output of prior activities. Offcourse I dont want to remove the email sending activity because it is in the prod environment and not developed by me. Is there any way to disable it? 回答1: You cannot disable one, but what you want to do is possible with the debug option in the editor. Just

How to get Azure Data Factory to Loop Through Files in a Folder

断了今生、忘了曾经 提交于 2019-12-23 04:33:16
问题 I am looking at the link below. https://azure.microsoft.com/en-us/updates/data-factory-supports-wildcard-file-filter-for-copy-activity/ We are supposed to have the ability to use wildcard characters in folder paths and file names. If we click on the 'Activity' and click 'Source', we see this view. I would like to loop through months any days, so it should be something like this view. Of course that doesn't actually work. I'm getting errors that read: ErrorCode: 'PathNotFound'. Message: 'The

Is possible to read an Azure Databricks table from Azure Data Factory?

三世轮回 提交于 2019-12-23 04:24:18
问题 I have a table into an Azure Databricks Cluster, i would like to replicate this data into an Azure SQL Database, to let another users analyze this data from Metabase. Is it possible to acess databricks tables through Azure Data factory? 回答1: No, unfortunately not. Databricks tables are typically temporary and last as long as your job/session is running. See here. You would need to persist your databricks table to some storage in order to access it. Change your databricks job to dump the table

How to setup custom Spark parameter in HDInsights cluster with Data Factory

谁说胖子不能爱 提交于 2019-12-22 18:17:29
问题 I am creating HDInsights cluster on Azure according to this desciption Now I would like to set up spark custom parameter, for example spark.yarn.appMasterEnv.PYSPARK3_PYTHON or spark_daemon_memory in time of cluster provisioning. Is it possible to setup using Data Factory/Automation Account? I can not find any example doing this. Thanks 回答1: You can use SparkConfig in Data Factory to pass these configurations to Spark. For example: "typeProperties": { ... "sparkConfig": { "spark.submit

Upload to ADLS from file stream

强颜欢笑 提交于 2019-12-21 20:52:39
问题 I am making a custom activity in ADF, which involves reading multiple files from Azure Storage Blobs, doing some work on them, and then finally writing a resulting file to the Azure Data Lake Store. Last step is where I stop, because as far as I can see, the .NET SDK only allows for uploading from a local file. Is there any way to to (programmatically) upload a file to ADL Store, where it is not from a local file? Could be a blob or a stream. If not, any workarounds? 回答1: Yes, it's possible

Azure Data Factory project with Visual Studio 2017

与世无争的帅哥 提交于 2019-12-20 17:47:16
问题 I am not sure whether Azure Data Factory project is supported on Visual Studio 2017 at the moment. I have just installed VS 2017 but cannot open our solution as there is one azure data factory project. Does Azure Data Factory support on Visual Studio 2017? 回答1: I tweeted the Visual Studio team with exactly this issue a couple of days ago. I even included a screen shot very similar to yours. This is the response I got: @mrpaulandrew We have no info about it yet. Our team is already working on