u-sql

Optimizer internal error while loading data from U-SQL table

天大地大妈咪最大 提交于 2019-12-10 23:15:13
问题 Is there a way to get around this error. "CQO: Internal Error - Optimizer internal error. Assert: a_drgcidChild->CLength() == UlSafeCLength(popMS->Pdrgcid()) in rlstreamset.cpp:499" Facing this issue while loading data from partitioned U-SQL table. @myData = SELECT * FROM dbo.MyTable; 回答1: If you encounter any system error message (or something that says Internal Error), please open a support ticket with us and/or send me your job link (if it happens on the cluster) or a self-contained

Transfer data from U-SQL managed table to Azure SQL Database table

风格不统一 提交于 2019-12-10 18:48:30
问题 I have a U-SQL managed table that contains schematized structured data. CREATE TABLE [AdlaDb].[dbo].[User] ( UserGuid Guid, Postcode string, Age int? DateOfBirth DateTime?, ) And a Azure SQL Database table. CREATE TABLE [SqlDb].[dbo].[User] ( UserGuid uniqueidentifier NOT NULL, Postcode varchar(15) NULL, Age int NULL, DateOfBirth Date NULL, ) I would like to transfer data from U-SQL managed table to Azure SQLDB table without losing the data types. I'm using azure data factory, seems like I

USQL - How To Select All Rows Between Two String Rows in USQL

两盒软妹~` 提交于 2019-12-10 18:17:34
问题 Here is my complete task description: I have to extract data from multiple files using u-sql and output it into csv file. Every input file contains multiple reports based on some string rows ("START OF ..." and "END OF ..." working as report separator). Here is an example (data format) of a single source (input) file : START OF DAILY ACCOUNT some data 1 some data 2 some data 3 some data n END OF DAILY ACCOUNT START OF LEDGER BALANCE some data 1 some data 2 some data 3 some data 4 some data 5

Should we delete DataLake Analytic Job after completion?

巧了我就是萌 提交于 2019-12-08 07:19:23
问题 We are submitting U-SQL jobs very frequently and we a see a list of jobs previously submitted in ADLA. We see the total storage utilization of Data Lake store is increasing day by day. All of our jobs submitted only update one single output file and size is around 10 MB. The current storage utilization of Data Lake store is 9.3 GB. We think it's due to the previous jobs resources are still saved in the Data Lake. Should we take care of this or we should do something here? 回答1: I think the job

Value too long failure when attempting to convert column data

本小妞迷上赌 提交于 2019-12-07 16:55:51
问题 Scenario I have a source file that contains blocks of JSON on each new line. I then have a simple U-SQL extract as follows where [RawString] represents each new line in the file and the [FileName] is defined as a variable from the @SourceFile path. @BaseExtract = EXTRACT [RawString] string, [FileName] string FROM @SourceFile USING Extractors.Text(delimiter:'\b', quoting : false); This executes without failure for the majority of my data and I'm able to parse the [RawString] as JSON further

U-SQL How can I get the current filename being processed to add to my extract output?

你说的曾经没有我的故事 提交于 2019-12-07 05:45:10
问题 I need to add meta data about the Row being processed. I need the filename to be added as a column. I looked at the ambulance demos in the Git repo, but can't figure out how to implement this. 回答1: You use a feature of U-SQL called 'file sets' and 'virtual columns'. In my simple example, I have two files in my input directory, I use file sets and refer to the virtual columns in the EXTRACT statement, eg // Filesets, file set with virtual column @q = EXTRACT rowId int, filename string,

Should we delete DataLake Analytic Job after completion?

∥☆過路亽.° 提交于 2019-12-06 14:46:22
We are submitting U-SQL jobs very frequently and we a see a list of jobs previously submitted in ADLA. We see the total storage utilization of Data Lake store is increasing day by day. All of our jobs submitted only update one single output file and size is around 10 MB. The current storage utilization of Data Lake store is 9.3 GB. We think it's due to the previous jobs resources are still saved in the Data Lake. Should we take care of this or we should do something here? I think the job data expires after a couple of weeks, but if you are concerned and do not need the data for auditing or

Handling Files With Different Columns in USQL

白昼怎懂夜的黑 提交于 2019-12-06 05:24:57
问题 I have a USQL script and CSV extractor to load my files. However some months the files may contain 4 columns and some months they may contain 5 columns. If I setup up my extractor with a column list for either 4 or 5 fields I get an error about the expected width of the file. Go check delimiters etc etc. No surprise. What is the work around to this problem please given USQL is still in a newbie and missing some basic error handling? I've tried using the silent clause in the extractor to

Data Lake Analytics U-SQL EXTRACT speed (Local vs Azure)

人盡茶涼 提交于 2019-12-06 02:08:06
Been looking into using the Azure Data Lake Analytics functionality to try and manipulate some Gzip’d xml data I have stored within Azures Blob Storage but I’m running into an interesting issue. Essentially when using U-SQL locally to process 500 of these xml files the processing time is extremely quick , roughly 40 seconds using 1 AU locally (which appears to be the limit). However when we run this same functionality from within Azure using 5 AU’s the processing takes 17+ minutes. We are eventually wanting to scale this up to ~ 20,000 files and more but have reduced the set to try and measure

Value too long failure when attempting to convert column data

爱⌒轻易说出口 提交于 2019-12-05 23:31:52
Scenario I have a source file that contains blocks of JSON on each new line. I then have a simple U-SQL extract as follows where [RawString] represents each new line in the file and the [FileName] is defined as a variable from the @SourceFile path. @BaseExtract = EXTRACT [RawString] string, [FileName] string FROM @SourceFile USING Extractors.Text(delimiter:'\b', quoting : false); This executes without failure for the majority of my data and I'm able to parse the [RawString] as JSON further down in my script without any problems. However, I seem to have an extra long row of data in a recent