pengine | s3fs | |
---|---|---|
2 | 7 | |
7 | 814 | |
- | 1.4% | |
0.0 | 7.8 | |
over 1 year ago | 22 days ago | |
Go | Python | |
BSD 2-clause "Simplified" License | BSD 3-clause "New" or "Revised" License |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
pengine
-
Mangle, a programming language for deductive database programming
Other resources for logic programming and Go:
ichiban/prolog - ISO Prolog interpreter in pure Go, getting close to v1: https://github.com/ichiban/prolog
trealla-prolog/go - ISO Prolog interpreter embedded via WASM: https://github.com/trealla-prolog/go
guregu/pengine - library for interfacing with Pengines (SWI-Prolog's RPC protocol): https://github.com/guregu/pengine
biscuit-auth/biscuit-go - Biscuits are a fancy auth token with a little Datalog engine: https://github.com/biscuit-auth/biscuit-go
I'm a big fan of logic programming. We've been seeing a small resurgence of interest in it (for example Yarn using Prolog made some waves) and I have some optimism for its future.
-
Prolog runtime for aws lambda?
This can work surprisingly well. There are some helpful predicates in guregu/predicates to deal with JSON and filesystem stuff (for example, you could hook up predicates.FS with something like jszwec/s3fs to easily load files from S3). If you end up needing a more mature Prolog, I also have a Go implementation of pengine_rpc/3 in the pengine package to easily call SWI.
s3fs
- Read files from s3 using Pandas/s3fs or AWS Data Wrangler?
- Gcsfuse: A user-space file system for interacting with Google Cloud Storage
-
what's the best python client for AWS automation these days?
- https://github.com/fsspec/s3fs (used by `pandas`, wraps aiobotocore)
- High-level, file-system like interface for S3 with AsyncIO support to replace/extend `boto3`
- Show HN: Query SQLite files stored in S3
-
Getting 403 return code from head_object even with s3:ListBucket permission
I'm using Python's s3fs library to check if a particular file exists in s3 with s3fs.S3FileSystem().exists(path), but I'm getting a Forbidden exception. From the stack trace, I can see it fails when calling s3's head_object method. The documentation for head_object method says:
What are some alternatives?
mangle
goofys - a high-performance, POSIX-ish Amazon S3 file system written in Go
prolog - The only reasonable scripting engine for Go.
rclone - "rsync for cloud storage" - Google Drive, S3, Dropbox, Backblaze B2, One Drive, Swift, Hubic, Wasabi, Google Cloud Storage, Azure Blob, Azure Files, Yandex Files
go - Trealla Prolog embedded in Go using WASM
s3www - Serve static files from any S3 compatible object storage services (Let's Encrypt ready)
dex-lang - Research language for array processing in the Haskell/ML family
aws-sdk-go-v2 - AWS SDK for the Go programming language.
biscuit-go
django-s3file - A lightweight file upload input for Django and Amazon S3
prolamb - SWI-Prolog lambda runtime
s3-proxy - S3 Reverse Proxy with GET, PUT and DELETE methods and authentication (OpenID Connect and Basic Auth)