q
csvq
Our great sponsors
q | csvq | |
---|---|---|
46 | 14 | |
10,106 | 1,446 | |
- | - | |
3.6 | 2.7 | |
3 months ago | 4 months ago | |
Python | Go | |
GNU General Public License v3.0 only | MIT License |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
q
-
I wrote this iCalendar (.ics) command-line utility to turn common calendar exports into more broadly compatible CSV files.
CSV utilities (still haven't pick a favorite one...): https://github.com/harelba/q https://github.com/BurntSushi/xsv https://github.com/wireservice/csvkit https://github.com/johnkerl/miller
- Segítség kérés Excel automatizáláshoz
-
Show HN: ClickHouse-local – a small tool for serverless data analytics
I think they're talking about https://github.com/harelba/q, which is not very fast.
-
sqly - execute SQL against CSV / JSON with shell
Apparently, there were many who thought the same thing; Tools to execute SQL against CSV were trdsql, q, csvq, TextQL. They were highly functional, hoewver, had many options and no input completion. I found it just a little difficult to use.
-
Q – Run SQL Directly on CSV or TSV Files
Hi, author of q here.
Regarding the error you got, q currently does not autodetect headers, so you'd need to add -H as a flag in order to use the "country" column name. You're absolutely correct on failing-fast here - It's a bug which i'll fix.
In general regarding speed - q supports automatic caching of the CSV files (through the "-C readwrite" flag). Once it's activated, it will write the data into another file (with a .qsql extension), and will use it automatically in further queries in order to speed things considerably.
Effectively, the .qsql files are regular sqlite3 files (with some metadata), and q can be used to query them directly (or any regular sqlite3 file), including the ability to seamlessly join between multiple sqlite3 files.
- PostgreSQL alternative for Large amounts of data
-
q VS trdsql - a user suggested alternative
2 projects | 25 Jun 2022
- One-liner for running queries against CSV files with SQLite
csvq
-
Fx – Terminal JSON Viewer
sure can do, if you already use that shell [1], but personally I like specific tools for specific jobs such as jq [2], fx, csvq [3] etc, there's value in decoupling shells from utils (modularity, speed, innovation etc).
[1] I don't but tempted to try, like its data-types concept
-
Tool to interact with CSV
csvq
-
Can SQL be used without an RDBMS?
There is a way of running SQL-like queries against CSV files.
-
Yq is a portable yq: command-line YAML, JSON, XML, CSV and properties processor
Lately I have had to do a lot of flat file analysis and tools along these lines have been a godsend. Will check this out.
My go to lately has been csvq (https://mithrandie.github.io/csvq/). Really nice to be able run complicated selects right over a CSV file with no setup at all.
-
Wie fusioniert man CSV tables?
csvq (https://mithrandie.github.io/csvq/)
-
Tool to explore big data sets
I usually do this with awk, my largest target files being half a TB in size for a project last year (and far too large to hold entirely in RAM). There are some other utilities like csvq and csvsql both of which let you write SQL-style queries against CSV files, but I'm not sure how they perform on large files. There's a nice list of CSV manipulation tools too if any of those jog your memory.
-
sqly - execute SQL against CSV / JSON with shell
Apparently, there were many who thought the same thing; Tools to execute SQL against CSV were trdsql, q, csvq, TextQL. They were highly functional, hoewver, had many options and no input completion. I found it just a little difficult to use.
- One-liner for running queries against CSV files with SQLite
-
Most efficient way to query .CSV files for Mac?
Please check out this tool https://github.com/mithrandie/csvq
-
Looking for: library to turn SQL (or abstracted) to code & execute against custom backend (slice of structs)
If you are looking to query nondb data with sql statements then you may want to check something like https://github.com/mithrandie/csvq (SQL for csv).
What are some alternatives?
textql - Execute SQL against structured text like CSV or TSV
querycsv - QueryCSV enables you to load CSV files and manipulate them using SQL queries then after you finish you can export the new values to a CSV file
octosql - OctoSQL is a query tool that allows you to join, analyse and transform data from multiple databases and file formats using SQL.
yq - yq is a portable command-line YAML, JSON, XML, CSV, TOML and properties processor
InquirerPy - :snake: Python port of Inquirer.js (A collection of common interactive command-line user interfaces)
yq - Command-line YAML, XML, TOML processor - jq wrapper for YAML/XML/TOML documents
xsv - A fast CSV command line toolkit written in Rust.
duckdb - DuckDB is an in-process SQL OLAP Database Management System
ledger - Double-entry accounting system with a command-line reporting interface
gsheet - gsheet is a CLI tool (and Golang package) for piping csv data to and from Google Sheets
simdjson - Parsing gigabytes of JSON per second : used by Facebook/Meta Velox, the Node.js runtime, ClickHouse, WatermelonDB, Apache Doris, Milvus, StarRocks
miller - Miller is like awk, sed, cut, join, and sort for name-indexed data such as CSV, TSV, and tabular JSON