police-settlements
disk.frame
police-settlements | disk.frame | |
---|---|---|
10 | 5 | |
143 | 593 | |
1.4% | 0.3% | |
0.0 | 0.0 | |
over 2 years ago | 3 months ago | |
R | R | |
- | GNU General Public License v3.0 or later |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
police-settlements
-
Biden Promised a Police Misconduct Database. He’s Yet to Deliver.
Traces of it can be found on github here.
- This is such a common tactic because police face ZERO accountability. The reporter was illegally arrested at a public park, they wanted to hide their actions from public view. The charges were dropped and the taxpayers will have to cover the lawsuit.
- Police misconduct settlements
- A FiveThirtyEight/The Marshall Project effort to collect comprehensive data on police misconduct settlements from 2010-19.
- Fivethirtyeight/police-settlements: comprehensive data on police misconduct
disk.frame
-
Do you code from memory? Or do you reference things?
Say hello to disk.frame.
- How can I read in only two columns from a massive 10+ GB tab file?
-
Data cleaning/ analysis 100-200 million rows of data. Is this doable in R, or is there another program I should try instead?
It depends on your hardware, but it should not be a problem. You might look into disk frame (https://diskframe.com) or similar packages.
-
is it possible to have my enviroment objects and work with them on my local drive instead of RAM?
If that doesn't work, the disk.frame package might help. It is new-ish and not common, but does seem to work with data on disk rather than in memory
-
We Test PCIe 4.0 Storage: The AnandTech 2021 SSD Benchmark Suite
> The speeds were just stunning to say the least at 15GB/s.
That is amazing. That is around DDR4-1866 speeds, and not far from DDR4-2666 (~21 GB/s). At those speeds I would happily work with dataframes sitting on the disk rather than in memory [1, 2]. Did you benchmark RAID 0 with less than four disks?
[1] R: https://github.com/xiaodaigh/disk.frame
What are some alternatives?
RedditExtractor - A minimalistic R wrapper for the Reddit API
db-benchmark - reproducible benchmark of database-like ops
opentripplanner - An R package to set up and use OpenTripPlanner (OTP) as a local or remote multimodal trip planner.
drake - An R-focused pipeline toolkit for reproducibility and high-performance computing
uaconflict_equipmentloss - This repo scrapes Oryxspioenkop (daily) to document and visualize equipment losses in the Russia-Ukraine war. https://www.oryxspioenkop.com/2022/02/attack-on-europe-documenting-equipment.html
r4ds - R for data science: a book
owidR - An R Package for Importing Data from Our World in Data
awesome-R - A curated list of awesome R packages, frameworks and software.
covid19-nor-data - Cleaned public data about Covid-19 in Norway
babynames - An R package containing US baby names from the SSA