amazon-s3

apache sling s3 bucket

随声附和 提交于 2020-02-25 05:41:26
问题 Within Apache Sling 11 is there a way to enable one of the folders within the content repository to be synchronized with an S3 bucket. Essentially allowing the use of multiple buckets if so desired or this a practice that should not be taken? 回答1: The biggest reason as to why the application was not creating any buckets within Amazon S3 as the server was starting was because I did not have the following installed within the server. Amazon OSGi SDK installed OAK Blob Cloud bundle Once these

Create EMR 5.3.0 with EMRFS (s3 bucket) as storage

假如想象 提交于 2020-02-25 05:28:12
问题 I'm trying to create EMR 5.3.0 with EMRFS (S3 bucket) as storage. Please provide your general guidance regarding this. Currently i'm using below command for creating EMR 5.3.0 with InstanceType=m4.2xlarge.Which is working fine, but with EMRFS as storage i'm not able to do aws emr create-cluster --name "DEMAPAUR001" --release-label emr-5.3.0 --service-role EMR_DefaultRole_Private --enable-debug --log-uri 's3n://xyz/trn' --ec2-attributes SubnetId=subnet-545e8823, KeyName=XXX --applications Name

How to avoid re-downloading media to S3 in Scrapy?

泄露秘密 提交于 2020-02-25 02:12:41
问题 I previously asked a similar question (How does Scrapy avoid re-downloading media that was downloaded recently?), but since I did not receive a definite answer I'll ask it again. I've downloaded a large number of files to an AWS S3 bucket using Scrapy's Files Pipeline. According to the documentation (https://doc.scrapy.org/en/latest/topics/media-pipeline.html#downloading-and-processing-files-and-images), this pipeline avoids "re-downloading media that was downloaded recently", but it does not

AWS Lambda and S3 - uploaded pdf file is blank/corrupt

守給你的承諾、 提交于 2020-02-25 02:08:47
问题 I have an Spring App(running on AWS Lambda) which gets a file and uploads it on AWS S3. The Spring Controller sends a MultipartFile to my method, where it's uploaded to AWS S3, using Amazon API Gateway. public static void uploadFile(MultipartFile mpFile, String fileName) throws IOException{ String dirPath = System.getProperty("java.io.tmpdir", "/tmp"); File file = new File(dirPath + "/" + fileName); OutputStream ops = new FileOutputStream(file); ops.write(mpFile.getBytes()); s3client

S3 Implementation for org.apache.parquet.io.InputFile?

自作多情 提交于 2020-02-24 12:07:55
问题 I am trying to write a Scala-based AWS Lambda to read Snappy compressed Parquet files based in S3. The process will write them backout in partitioned JSON files. I have been trying to use the org.apache.parquet.hadoop.ParquetFileReader class to read the files... the non-deprecated way to do this appears to pass it a implementation of the org.apache.parquet.io.InputFile interface. There is one for Hadoop (HadoopInputFile)... but I cannot find one for S3. I also tried some of the deprecated

S3 Implementation for org.apache.parquet.io.InputFile?

时光毁灭记忆、已成空白 提交于 2020-02-24 12:05:10
问题 I am trying to write a Scala-based AWS Lambda to read Snappy compressed Parquet files based in S3. The process will write them backout in partitioned JSON files. I have been trying to use the org.apache.parquet.hadoop.ParquetFileReader class to read the files... the non-deprecated way to do this appears to pass it a implementation of the org.apache.parquet.io.InputFile interface. There is one for Hadoop (HadoopInputFile)... but I cannot find one for S3. I also tried some of the deprecated

How do I unlock or delete an AWS S3 bucket that inadvertently locked with a bucket policy?

天涯浪子 提交于 2020-02-24 11:15:14
问题 I am the AWS account owner, admin (both IAM and root password - not recommended I know, but have some legacy apps that depend on root pw). I recently added a bucket policy which has subsequently rendered the bucket locked an inaccessible to me: $ aws s3 rb s3://<my-bucket> --force A client error (AccessDenied) occurred when calling the ListObjects operation: Access Denied Completed 1 part(s) with ... file(s) remaining remove_bucket failed: s3://<my-bucket>/ A client error (AccessDenied)

Java Heap Space is insufficient to upload files on AWS S3

柔情痞子 提交于 2020-02-24 09:09:34
问题 I'm trying to upload a file on AWS S3 by using Java-AWS API. The problem is my application is unable to upload large sized files because the heap is reaching its limit. Error: java.lang.OutOfMemoryError: Java heap space I personally think extending heap memory isn't a permanent solution because I have to upload file upto 100 gb. What should I do ? Here is the code snippet: BasicAWSCredentials awsCreds = new BasicAWSCredentials(AID, Akey); AmazonS3 s3Client = AmazonS3ClientBuilder.standard()

Java Heap Space is insufficient to upload files on AWS S3

巧了我就是萌 提交于 2020-02-24 09:07:56
问题 I'm trying to upload a file on AWS S3 by using Java-AWS API. The problem is my application is unable to upload large sized files because the heap is reaching its limit. Error: java.lang.OutOfMemoryError: Java heap space I personally think extending heap memory isn't a permanent solution because I have to upload file upto 100 gb. What should I do ? Here is the code snippet: BasicAWSCredentials awsCreds = new BasicAWSCredentials(AID, Akey); AmazonS3 s3Client = AmazonS3ClientBuilder.standard()

Serve static files in Flask from private AWS S3 bucket

痞子三分冷 提交于 2020-02-23 11:48:05
问题 I am developing a Flask app running on Heroku that allows users to upload images. The app has a page displaying the user's images in a table. For developing purposes, I am saving the uploaded files to Heroku's ephemeral file system, and everything works fine: the images are correctly loaded and displayed (I am using the last method shown here implying the use of send_from_directory()). Now I have moved the storage to S3 and I am trying to adapt the code. I use boto3 to upload the files to the