azure-data-factory

Azure Data Factory Copy Activity - Can column mappings be composites or use a Templating function?

喜你入骨 提交于 2019-12-06 10:08:36
问题 When using a Data Factory Copy Activity to take delimited text files from Blob Storage to Table Storage, I am able to specify the input and output schemas and the relevant one-to-one Column Mappings. It is possible to concatenate or format one or more columns (with $$Text.Format perhaps?) in the source schema and map these to a single column in the destination? If possible, this would be especially useful if I can also set the Partition Key using a similar approach (a composite key taken from

Bringing incremental data in from REST APIs into SQL azure

你说的曾经没有我的故事 提交于 2019-12-06 03:23:29
My needs are following: - Need to fetch data from a 3rd party API into SQL azure. The API's will be queried everyday for incremental data and may require pagination as by default any API response will give only Top N records. The API also needs an auth token to work, which is the first call before we start downloading data from endpoints. Due to last two reasons, I've opted for Function App which will be triggered daily rather than data factory which can query web APIs. Is there a better way to do this? Also I am thinking of pushing all JSON into Blob store and then parsing data from the JSON

Azure Data Factory Test Framework

老子叫甜甜 提交于 2019-12-05 21:04:05
Are there any automatic testing mechanism available for azure data factory pipelines? Does the azure data factory visual studio project come with any test suite of its own? Any help highly appreciated Thanks Not that I'm aware of, but happy to be told otherwise. I suggest you post this on Microsoft's user voice page as a feedback idea. Then people searching will come here, go to that link and vote to get something developed. https://feedback.azure.com/forums/270578-data-factory/filters/my_feedback?query=Unit%20Testing%20for%20ADF%20Projects Hope this helps. EDIT after comment: You could use a

How to read files with .xlsx and .xls extension in Azure data factory?

与世无争的帅哥 提交于 2019-12-05 08:48:52
I am trying to read and excel file in Azure Blob Storage with .xlsx extension in my azure data factory dataset. it throws following error Error found when processing 'Csv/Tsv Format Text' source 'Filename.xlsx' with row number 3: found more columns than expected column count: 1. What are the right Column and row delimiters for excel files to be read in azure Data factory Excel files have a proprietary format and are not simple delimited files. As indicated here , Azure Data Factory does not have a direct option to import Excel files, eg you cannot create a Linked Service to an Excel file and

Parse json file in U-SQL

一世执手 提交于 2019-12-05 04:55:51
I'm trying to parse below Json file using USQL but keep getting error. Json file@ {"dimBetType_SKey":1,"BetType_BKey":1,"BetTypeName":"Test1"} {"dimBetType_SKey":2,"BetType_BKey":2,"BetTypeName":"Test2"} {"dimBetType_SKey":3,"BetType_BKey":3,"BetTypeName":"Test3"} Below is the USQL script, I'm trying to extract the data from above file. REFERENCE ASSEMBLY [Newtonsoft.Json]; REFERENCE ASSEMBLY [Microsoft.Analytics.Samples.Formats]; DECLARE @Full_Path string = "adl://xxxx.azuredatalakestore.net/2017/03/28/00_0_66ffdd26541742fab57139e95080e704.json"; DECLARE @Output_Path = "adl://xxxx

Azure Data Factory - How can I trigger Scheduled/OneTime Pipelines?

扶醉桌前 提交于 2019-12-05 03:05:17
Background : I have scheduled pipelines running for copying data from source to destination. This is scheduled to run daily at a specific time. Problem : The input dataset to the pipeline is external and not available at specific time intervals. This means the copy activity will have to wait until the Scheduled Start time mentioned in the Pipeline to kickoff. Considering the volume of data, I don't want to waste my time here. Requirement : At any given time I have access to the time when my input data set is available. With this in hand, I want to know how to trigger a ADF Pipeline from C#

How to authenticate with Azure Analysis Services from Azure batch and data factory

梦想的初衷 提交于 2019-12-04 17:51:38
I have a c sharp class library that connects to Azure Analysis Services using the AMO library. I'd like to use this as part of my data factory pipeline to refresh cube partitions. This is done through Azure batch as a custom .net activity. var server = new Server(); server.Connect("Provider=MSOLAP;Data Source=asazure://uksouth.asazure.windows.net/abcd;Initial Catalog=xyz"); Running this locally works fine, however this will not run in the cloud. It currently errors out as it is not being run under my user account. I know that I can add a username and password to the connection string, but I

Azure Data Factory Copy Activity - Can column mappings be composites or use a Templating function?

对着背影说爱祢 提交于 2019-12-04 17:02:58
When using a Data Factory Copy Activity to take delimited text files from Blob Storage to Table Storage, I am able to specify the input and output schemas and the relevant one-to-one Column Mappings. It is possible to concatenate or format one or more columns (with $$Text.Format perhaps?) in the source schema and map these to a single column in the destination? If possible, this would be especially useful if I can also set the Partition Key using a similar approach (a composite key taken from 2 or more source column values or better still, the SliceStart plus a concatenated source column value

Upload to ADLS from file stream

二次信任 提交于 2019-12-04 11:43:36
I am making a custom activity in ADF, which involves reading multiple files from Azure Storage Blobs, doing some work on them, and then finally writing a resulting file to the Azure Data Lake Store. Last step is where I stop, because as far as I can see, the .NET SDK only allows for uploading from a local file. Is there any way to to (programmatically) upload a file to ADL Store, where it is not from a local file? Could be a blob or a stream. If not, any workarounds? Yes, it's possible to upload from Stream, the trick is to create file first and then append your stream to it: string

'Set-AzureRmDataFactoryV2' is not recognized as the name of a cmdlet

爱⌒轻易说出口 提交于 2019-12-04 04:49:00
问题 I have installed Powershell 6.0.0 on Ubuntu 16.04. I am getting following error when trying to create Azure Data Factory using Powershell Set-AzureRmDataFactoryV2 : The term 'Set-AzureRmDataFactoryV2' is not recognized as the name of a cmdlet, function, script file, or operable program. Check the spelling of the name, or if a path was included, verify that the path is correct and try again. At line:1 char:16 + $DataFactory = Set-AzureRmDataFactoryV2 -ResourceGroupName $ResGrp.Re ... + ~~~~~~~