amazon-s3

Unable to execute HTTP request: Timeout waiting for connection from pool in Flink

≯℡__Kan透↙ 提交于 2021-01-29 09:02:26
问题 I'm working on an app which uploads some files to an s3 bucket and at a later point, it reads files from s3 bucket and pushes it to my database . I'm using Flink 1.4.2 and fs.s3a API for reading and write files from the s3 bucket. Uploading files to s3 bucket works fine without any problem but when the second phase of my app that is reading those uploaded files from s3 starts, my app is throwing following error : Caused by: java.io.InterruptedIOException: Reopen at position 0 on s3a:/

dynamically folder creation in s3 bucket from pyspark job

与世无争的帅哥 提交于 2021-01-29 09:01:37
问题 I am writing data into s3 bucket and creating parquet files using pyspark . MY bucket structure looks like below: s3a://rootfolder/subfolder/table/ subfolder and table these two folders should be created at run time if folders do not exist , and if folders exist parquet files should inside folder table . when I am running pyspark program from local machine it creates extra folder with _$folder$ (like table_$folder$ ) but if same program is run from emr it creates with _SUCCESS . writing into

S3 Python Download with Progress Bar

≯℡__Kan透↙ 提交于 2021-01-29 08:55:09
问题 Couldn't comment on the initial thread where I adapted this code (Track download progress of S3 file using boto3 and callbacks) so hopefully someone can help me here. I was able to use this code to show a progress bar for file uploads, now I need the do the same thing for file downloads from AWS S3. Any help would be GREATLY APPRECIATED! I know I need to get the size of the file from S3 instead of from the local file system. I'm sure there is some silly code I need to adjust to make this work

FineUploader with S3 how to specify object name on S3

Deadly 提交于 2021-01-29 08:24:27
问题 Where do I specify the format or pattern for the server name that FineUploader creates when uploading to Amazon S3? I need to have greater control over how the file name and path is created on S3. Thanks. 回答1: Use the objectProperties.key option. You can provide a String value or a function that returns the intended key as a String . The following example demonstrates setting a key and performing a blocking action ( $.post ). It uses the jQuery plugin, // <snip> objectProperties: { key:

PHP + FFMPEG + S3. Transcode video directly between S3 buckets

眉间皱痕 提交于 2021-01-29 08:20:39
问题 I have: S3 INPUT bucket S3 OUTPUT bucket PHP ffmpeg Is it possible to read file directly from INPUT bucket → Transcode to another format → Save into OUTPUT bucket. Please let me know manuals, libraries, frameworks, any stuff that helps to understand how to do it. At least python realisations also welcome. At least at least some different language also OK. Input file size may be more than 10Gb so writing whole file in RAM is undesirable. Some chunk-based way is preferable. Output format is

Spark Scala S3 storage: permission denied

|▌冷眼眸甩不掉的悲伤 提交于 2021-01-29 08:12:33
问题 I've read a lot of topic on Internet on how to get working Spark with S3 still there's nothing working properly. I've downloaded : Spark 2.3.2 with hadoop 2.7 and above. I've copied only some libraries from Hadoop 2.7.7 (which matches Spark/Hadoop version) to Spark jars folder: hadoop-aws-2.7.7.jar hadoop-auth-2.7.7.jar aws-java-sdk-1.7.4.jar Still I can't use nor S3N nor S3A to get my file read by spark: For S3A I have this exception: sc.hadoopConfiguration.set("fs.s3a.access.key",

Unable to fetch complete records from Salesforce using Python

安稳与你 提交于 2021-01-29 07:54:41
问题 I am trying to fetch the data from salesforce using the simple_salesforce library in python. I am able to get the correct count of records while running the count query. But while I am trying to put that results (in the form of list) into s3 as a JSON object, not as many reocrds are getting persisted as I captured from Salesforce. Here is the piece of code: result = sf.query("SELECT ID FROM Opportunity")['records'] object.put(Body=(bytes(json.dumps(result, indent=2).encode('UTF-8')))) Is the

ASP NET Core AWS No RegionEndpoint or ServiceURL configured when deployed to Heroku

冷暖自知 提交于 2021-01-29 06:07:07
问题 Hi I deployed my WEB API project to Heroku but I get an AWS exception 'No RegionEndpoint or ServiceURL configured' I tried to add 'region' and 'aws_default_region' to heroku's app enviromental variables and set them to 'eu-north-1' but it didn't help. Also I added these lines to my web.config file under configuration but I still get that exception: <appSettings> <add key="AWSRegion" value="eu-north-1"/> </appSettings> <aws region="eu-north-1"/> I can't create my own aws client and set the

Training Failed - AWS Machine Learning

浪子不回头ぞ 提交于 2021-01-29 05:45:48
问题 I am working on Aws Machine learning with MERN(Mongodb,Express,React,NodeJS)Stack Code.But the issue is that when I upload the data file (.csv file) for process machine learning after sometime process training is failed with TrainingFailed Error which is follow: AlgorithmError: CannotStartContainerError. Please make sure the container can be run with 'docker run train'. Please refer SageMaker documentation for details. It is possible that the Dockerfile's entrypoint is not properly defined,

AWS Transcribe S3 .wav file to text

↘锁芯ラ 提交于 2021-01-29 05:39:02
问题 I am trying to use aws Transcribe to convert a wav file to text. I have uploaded a wav file to S3, which is located here and it has public read/write permissions: https://s3.us-east-1.amazonaws.com/csld8xmsdksdf8s9sk3mmdjsdifkjksdijsldk/Transcribe2.wav. The wav file is valid. I can download it in my browser and replay it (and it sounds like the origin recording), so I think we can rule out an invalid input file, file permissions, etc. I am using java version: 1.8.0_275 for mac. I expect my