google-bigquery

SPLIT key-value-pairs to Columns in Google BigQuery

血红的双手。 提交于 2021-02-10 06:56:12
问题 I am quite new to Google BigQuery and definitely struggling. My table has the following content: +----------+----------------------------------------+ | order_id | line_items | +----------+----------------------------------------+ | 123 | id:1|qy:1|sum:1.00;id:2|qy:6|sum:4.50; | +----------+----------------------------------------+ | 456 | id:1|qy:3|sum:3.00;id:3|qy:4|sum:3.20; | +----------+----------------------------------------+ I would need to look it like this: +----------+----+----+---

Google App Script: Javascript runtime exited unexpectedly

▼魔方 西西 提交于 2021-02-10 06:14:43
问题 I have written an app script that will interact with the bigQuery API and fetch the bigQuery results. API is been hit around close to 60-70 times. In this process sometimes I'm getting the following error message: "JavaScript runtime exited unexpectedly" could someone help me here? Let me know if you need any additional information function fetchTables() { var timezone = "GMT+" + new Date().getTimezoneOffset()/60; var date = Utilities.formatDate(new Date(), timezone, "yyyy-MM-dd HH:mm");

Why is my export sink from Stackdriver only loading the latest audit logs into BigQuery and no historical?

我与影子孤独终老i 提交于 2021-02-10 05:51:31
问题 I created an export sink in Stackdriver to load audit logs into BigQuery. I want to be able to see audit logs from the past 3 months. However, when I queried the tables in BigQuery, I am only seeing logs from today and no earlier. I applied the following filters to my export sink. I also tried removing the timestamp filter but still only seeing logs from today and no prior. resource.type="bigquery_dataset" timestamp > "2019-05-01T23:59:09.739Z" 回答1: Exports only work for new entries. Per the

most reliable format for large bigquery load jobs

[亡魂溺海] 提交于 2021-02-09 09:13:37
问题 I have a 100 GB table that I'm trying to load into google bigquery. It is stored as a single 100GB avro file on GCS. Currently my bq load job is failing with an unhelpful error message: UDF worker timed out during execution.; Unexpected abort triggered for worker avro-worker-156907: request_timeout I'm thinking of trying a different format. I understand that bigquery supports several formats (AVRO, JSON, CSV, Parquet, etc) and that in principle one can load large datasets in any of these

most reliable format for large bigquery load jobs

萝らか妹 提交于 2021-02-09 09:12:17
问题 I have a 100 GB table that I'm trying to load into google bigquery. It is stored as a single 100GB avro file on GCS. Currently my bq load job is failing with an unhelpful error message: UDF worker timed out during execution.; Unexpected abort triggered for worker avro-worker-156907: request_timeout I'm thinking of trying a different format. I understand that bigquery supports several formats (AVRO, JSON, CSV, Parquet, etc) and that in principle one can load large datasets in any of these

most reliable format for large bigquery load jobs

和自甴很熟 提交于 2021-02-09 09:11:13
问题 I have a 100 GB table that I'm trying to load into google bigquery. It is stored as a single 100GB avro file on GCS. Currently my bq load job is failing with an unhelpful error message: UDF worker timed out during execution.; Unexpected abort triggered for worker avro-worker-156907: request_timeout I'm thinking of trying a different format. I understand that bigquery supports several formats (AVRO, JSON, CSV, Parquet, etc) and that in principle one can load large datasets in any of these

most reliable format for large bigquery load jobs

六眼飞鱼酱① 提交于 2021-02-09 09:11:05
问题 I have a 100 GB table that I'm trying to load into google bigquery. It is stored as a single 100GB avro file on GCS. Currently my bq load job is failing with an unhelpful error message: UDF worker timed out during execution.; Unexpected abort triggered for worker avro-worker-156907: request_timeout I'm thinking of trying a different format. I understand that bigquery supports several formats (AVRO, JSON, CSV, Parquet, etc) and that in principle one can load large datasets in any of these

Correlated sub-query causing this error - LEFT OUTER JOIN cannot be used without a condition that is an equality of fields from both sides of the join

浪子不回头ぞ 提交于 2021-02-08 23:45:42
问题 When writing correlated sub-queries in BigQuery using a WHERE clause, it is resulting in an Error. LEFT OUTER JOIN cannot be used without a condition that is an equality of fields from both sides of the join. However it works in other SQL engines, I am not understanding the error or how to rewrite the queries. Different queries of the same type causing the same error. SELECT out.* FROM `august-monolith-205810.Assignment.students` AS out WHERE 2 >= ( SELECT COUNT(*) FROM `august-monolith

Correlated sub-query causing this error - LEFT OUTER JOIN cannot be used without a condition that is an equality of fields from both sides of the join

邮差的信 提交于 2021-02-08 23:45:10
问题 When writing correlated sub-queries in BigQuery using a WHERE clause, it is resulting in an Error. LEFT OUTER JOIN cannot be used without a condition that is an equality of fields from both sides of the join. However it works in other SQL engines, I am not understanding the error or how to rewrite the queries. Different queries of the same type causing the same error. SELECT out.* FROM `august-monolith-205810.Assignment.students` AS out WHERE 2 >= ( SELECT COUNT(*) FROM `august-monolith

import pyarrow not working <- error is “ValueError: The pyarrow library is not installed, please install pyarrow to use the to_arrow() function.”

[亡魂溺海] 提交于 2021-02-08 15:47:12
问题 I have tried installing it in the terminal and in juypter lab and it says that it has been successfully installed but when I run df = query_job.to_dataframe() I keep getting the error " ValueError: The pyarrow library is not installed, please install pyarrow to use the to_arrow() function.". I have no idea how to fix this. Any advice? I am trying to access data from google data studio ultimately with the code, from google.cloud import bigquery import pandas import numpy import pyarrow